2026-03-10T11:25:39.132 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-10T11:25:39.136 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T11:25:39.155 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1016 branch: squid description: orch/cephadm/with-work/{0-distro/centos_9.stream fixed-2 mode/packaged mon_election/classic msgr/async start tasks/rados_python} email: null first_in_suite: false flavor: default job_id: '1016' last_in_suite: false machine_type: vps name: kyr-2026-03-10_01:00:38-orch-squid-none-default-vps no_nested_subset: false openstack: - volumes: count: 4 size: 10 os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon election default strategy: 1 ms type: async mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 osd shutdown pgref assert: true flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - but it is still running - overall HEALTH_ - \(OSDMAP_FLAGS\) - \(PG_ - \(OSD_ - \(OBJECT_ - \(POOL_APP_NOT_ENABLED\) log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} cephadm: cephadm_mode: cephadm-package install: ceph: extra_system_packages: deb: - python3-pytest rpm: - python3-pytest flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_packages: - cephadm extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 - scontext=system_u:system_r:getty_t:s0 workunit: branch: tt-squid sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - mon.a - mon.c - mgr.y - osd.0 - osd.1 - osd.2 - osd.3 - client.0 - ceph.rgw.foo.a - node-exporter.a - alertmanager.a - - mon.b - mgr.x - osd.4 - osd.5 - osd.6 - osd.7 - client.1 - prometheus.a - grafana.a - node-exporter.b - ceph.iscsi.iscsi.a seed: 8043 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b targets: vm06.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCXGEs2B9UKkPUAKYhbltC6ExG1fExjQQ8zUdzw/roCf8VYpbZJl3tiHcYmgh6udhisAFIi4mDsUVZCJS/5GAI4= vm09.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPBavJNCpl2BMFC7Ue+GHu1ARdT2hiRPMxU26OZldTyStp6r5+vz4EisNq1akv0Ig5c1mp04+i9BTvdxiTtuAAE= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install nvmetcli nvme-cli -y - install: null - cephadm: conf: mgr: debug mgr: 20 debug ms: 1 - workunit: clients: client.0: - rados/test_python.sh timeout: 1h teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-10_01:00:38 tube: vps user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 2026-03-10T11:25:39.155 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa; will attempt to use it 2026-03-10T11:25:39.155 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa/tasks 2026-03-10T11:25:39.155 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-10T11:25:39.155 INFO:teuthology.task.internal:Checking packages... 2026-03-10T11:25:39.156 INFO:teuthology.task.internal:Checking packages for os_type 'centos', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-10T11:25:39.156 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-10T11:25:39.156 INFO:teuthology.packaging:ref: None 2026-03-10T11:25:39.156 INFO:teuthology.packaging:tag: None 2026-03-10T11:25:39.156 INFO:teuthology.packaging:branch: squid 2026-03-10T11:25:39.156 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T11:25:39.156 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&ref=squid 2026-03-10T11:25:39.945 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678.ge911bdeb 2026-03-10T11:25:39.946 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-10T11:25:39.947 INFO:teuthology.task.internal:no buildpackages task found 2026-03-10T11:25:39.947 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-10T11:25:39.947 INFO:teuthology.task.internal:Saving configuration 2026-03-10T11:25:39.951 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-10T11:25:39.952 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-10T11:25:39.960 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm06.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1016', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 11:24:25.365545', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:06', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCXGEs2B9UKkPUAKYhbltC6ExG1fExjQQ8zUdzw/roCf8VYpbZJl3tiHcYmgh6udhisAFIi4mDsUVZCJS/5GAI4='} 2026-03-10T11:25:39.966 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm09.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1016', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 11:24:25.365941', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:09', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPBavJNCpl2BMFC7Ue+GHu1ARdT2hiRPMxU26OZldTyStp6r5+vz4EisNq1akv0Ig5c1mp04+i9BTvdxiTtuAAE='} 2026-03-10T11:25:39.966 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-10T11:25:39.967 INFO:teuthology.task.internal:roles: ubuntu@vm06.local - ['mon.a', 'mon.c', 'mgr.y', 'osd.0', 'osd.1', 'osd.2', 'osd.3', 'client.0', 'ceph.rgw.foo.a', 'node-exporter.a', 'alertmanager.a'] 2026-03-10T11:25:39.967 INFO:teuthology.task.internal:roles: ubuntu@vm09.local - ['mon.b', 'mgr.x', 'osd.4', 'osd.5', 'osd.6', 'osd.7', 'client.1', 'prometheus.a', 'grafana.a', 'node-exporter.b', 'ceph.iscsi.iscsi.a'] 2026-03-10T11:25:39.967 INFO:teuthology.run_tasks:Running task console_log... 2026-03-10T11:25:39.974 DEBUG:teuthology.task.console_log:vm06 does not support IPMI; excluding 2026-03-10T11:25:39.980 DEBUG:teuthology.task.console_log:vm09 does not support IPMI; excluding 2026-03-10T11:25:39.980 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f928b572170>, signals=[15]) 2026-03-10T11:25:39.980 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-10T11:25:39.981 INFO:teuthology.task.internal:Opening connections... 2026-03-10T11:25:39.981 DEBUG:teuthology.task.internal:connecting to ubuntu@vm06.local 2026-03-10T11:25:39.982 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T11:25:40.043 DEBUG:teuthology.task.internal:connecting to ubuntu@vm09.local 2026-03-10T11:25:40.043 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T11:25:40.101 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-10T11:25:40.102 DEBUG:teuthology.orchestra.run.vm06:> uname -m 2026-03-10T11:25:40.158 INFO:teuthology.orchestra.run.vm06.stdout:x86_64 2026-03-10T11:25:40.159 DEBUG:teuthology.orchestra.run.vm06:> cat /etc/os-release 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:NAME="CentOS Stream" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:VERSION="9" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:ID="centos" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:ID_LIKE="rhel fedora" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_ID="9" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:PLATFORM_ID="platform:el9" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:ANSI_COLOR="0;31" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:LOGO="fedora-logo-icon" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:HOME_URL="https://centos.org/" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T11:25:40.215 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T11:25:40.216 INFO:teuthology.lock.ops:Updating vm06.local on lock server 2026-03-10T11:25:40.220 DEBUG:teuthology.orchestra.run.vm09:> uname -m 2026-03-10T11:25:40.234 INFO:teuthology.orchestra.run.vm09.stdout:x86_64 2026-03-10T11:25:40.235 DEBUG:teuthology.orchestra.run.vm09:> cat /etc/os-release 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:NAME="CentOS Stream" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:VERSION="9" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:ID="centos" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:ID_LIKE="rhel fedora" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:VERSION_ID="9" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:PLATFORM_ID="platform:el9" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:ANSI_COLOR="0;31" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:LOGO="fedora-logo-icon" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:HOME_URL="https://centos.org/" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T11:25:40.290 INFO:teuthology.orchestra.run.vm09.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T11:25:40.291 INFO:teuthology.lock.ops:Updating vm09.local on lock server 2026-03-10T11:25:40.295 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-10T11:25:40.297 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-10T11:25:40.297 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-10T11:25:40.298 DEBUG:teuthology.orchestra.run.vm06:> test '!' -e /home/ubuntu/cephtest 2026-03-10T11:25:40.300 DEBUG:teuthology.orchestra.run.vm09:> test '!' -e /home/ubuntu/cephtest 2026-03-10T11:25:40.348 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-10T11:25:40.349 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-10T11:25:40.350 DEBUG:teuthology.orchestra.run.vm06:> test -z $(ls -A /var/lib/ceph) 2026-03-10T11:25:40.355 DEBUG:teuthology.orchestra.run.vm09:> test -z $(ls -A /var/lib/ceph) 2026-03-10T11:25:40.370 INFO:teuthology.orchestra.run.vm06.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T11:25:40.407 INFO:teuthology.orchestra.run.vm09.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T11:25:40.407 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-10T11:25:40.415 DEBUG:teuthology.orchestra.run.vm06:> test -e /ceph-qa-ready 2026-03-10T11:25:40.433 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T11:25:40.632 DEBUG:teuthology.orchestra.run.vm09:> test -e /ceph-qa-ready 2026-03-10T11:25:40.646 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T11:25:40.816 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-10T11:25:40.817 INFO:teuthology.task.internal:Creating test directory... 2026-03-10T11:25:40.817 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T11:25:40.819 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T11:25:40.834 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-10T11:25:40.835 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-10T11:25:40.836 INFO:teuthology.task.internal:Creating archive directory... 2026-03-10T11:25:40.836 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T11:25:40.876 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T11:25:40.893 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-10T11:25:40.894 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-10T11:25:40.894 DEBUG:teuthology.orchestra.run.vm06:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T11:25:40.945 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T11:25:40.945 DEBUG:teuthology.orchestra.run.vm09:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T11:25:40.960 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T11:25:40.960 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T11:25:40.988 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T11:25:41.010 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T11:25:41.020 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T11:25:41.025 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T11:25:41.034 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T11:25:41.036 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-10T11:25:41.037 INFO:teuthology.task.internal:Configuring sudo... 2026-03-10T11:25:41.037 DEBUG:teuthology.orchestra.run.vm06:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T11:25:41.064 DEBUG:teuthology.orchestra.run.vm09:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T11:25:41.101 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-10T11:25:41.103 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-10T11:25:41.103 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T11:25:41.131 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T11:25:41.155 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T11:25:41.211 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T11:25:41.271 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:25:41.271 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T11:25:41.330 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T11:25:41.353 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T11:25:41.408 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:25:41.408 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T11:25:41.466 DEBUG:teuthology.orchestra.run.vm06:> sudo service rsyslog restart 2026-03-10T11:25:41.468 DEBUG:teuthology.orchestra.run.vm09:> sudo service rsyslog restart 2026-03-10T11:25:41.494 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T11:25:41.532 INFO:teuthology.orchestra.run.vm09.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T11:25:41.803 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-10T11:25:41.805 INFO:teuthology.task.internal:Starting timer... 2026-03-10T11:25:41.805 INFO:teuthology.run_tasks:Running task pcp... 2026-03-10T11:25:41.808 INFO:teuthology.run_tasks:Running task selinux... 2026-03-10T11:25:41.810 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0', 'scontext=system_u:system_r:getty_t:s0']} 2026-03-10T11:25:41.810 INFO:teuthology.task.selinux:Excluding vm06: VMs are not yet supported 2026-03-10T11:25:41.810 INFO:teuthology.task.selinux:Excluding vm09: VMs are not yet supported 2026-03-10T11:25:41.810 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-10T11:25:41.810 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-10T11:25:41.810 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-10T11:25:41.810 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-10T11:25:41.811 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-10T11:25:41.812 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-10T11:25:41.813 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-10T11:25:42.421 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-10T11:25:42.426 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-10T11:25:42.427 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryymlqwkns --limit vm06.local,vm09.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-10T11:27:44.536 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm06.local'), Remote(name='ubuntu@vm09.local')] 2026-03-10T11:27:44.537 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm06.local' 2026-03-10T11:27:44.537 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T11:27:44.600 DEBUG:teuthology.orchestra.run.vm06:> true 2026-03-10T11:27:44.674 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm06.local' 2026-03-10T11:27:44.674 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm09.local' 2026-03-10T11:27:44.675 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T11:27:44.738 DEBUG:teuthology.orchestra.run.vm09:> true 2026-03-10T11:27:44.814 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm09.local' 2026-03-10T11:27:44.814 INFO:teuthology.run_tasks:Running task clock... 2026-03-10T11:27:44.817 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-10T11:27:44.817 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T11:27:44.817 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T11:27:44.820 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T11:27:44.820 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T11:27:44.851 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T11:27:44.867 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T11:27:44.894 INFO:teuthology.orchestra.run.vm06.stderr:sudo: ntpd: command not found 2026-03-10T11:27:44.896 INFO:teuthology.orchestra.run.vm09.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T11:27:44.905 INFO:teuthology.orchestra.run.vm06.stdout:506 Cannot talk to daemon 2026-03-10T11:27:44.914 INFO:teuthology.orchestra.run.vm09.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T11:27:44.919 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T11:27:44.934 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T11:27:44.944 INFO:teuthology.orchestra.run.vm09.stderr:sudo: ntpd: command not found 2026-03-10T11:27:44.962 INFO:teuthology.orchestra.run.vm09.stdout:506 Cannot talk to daemon 2026-03-10T11:27:44.977 INFO:teuthology.orchestra.run.vm09.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T11:27:44.980 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-10T11:27:44.984 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T11:27:44.984 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-10T11:27:44.984 INFO:teuthology.orchestra.run.vm06.stdout:^? fa.gnudb.org 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T11:27:44.984 INFO:teuthology.orchestra.run.vm06.stdout:^? 47.ip-51-75-67.eu 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T11:27:44.984 INFO:teuthology.orchestra.run.vm06.stdout:^? node-1.infogral.is 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T11:27:44.984 INFO:teuthology.orchestra.run.vm06.stdout:^? ntp2.uni-ulm.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T11:27:44.994 INFO:teuthology.orchestra.run.vm09.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T11:27:45.042 INFO:teuthology.orchestra.run.vm09.stderr:bash: line 1: ntpq: command not found 2026-03-10T11:27:45.045 INFO:teuthology.orchestra.run.vm09.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T11:27:45.045 INFO:teuthology.orchestra.run.vm09.stdout:=============================================================================== 2026-03-10T11:27:45.045 INFO:teuthology.orchestra.run.vm09.stdout:^? node-1.infogral.is 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T11:27:45.045 INFO:teuthology.orchestra.run.vm09.stdout:^? ntp2.uni-ulm.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T11:27:45.045 INFO:teuthology.orchestra.run.vm09.stdout:^? fa.gnudb.org 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T11:27:45.045 INFO:teuthology.orchestra.run.vm09.stdout:^? 47.ip-51-75-67.eu 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T11:27:45.046 INFO:teuthology.run_tasks:Running task pexec... 2026-03-10T11:27:45.048 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-10T11:27:45.048 DEBUG:teuthology.orchestra.run.vm06:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T11:27:45.048 DEBUG:teuthology.orchestra.run.vm09:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T11:27:45.051 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf remove nvme-cli -y 2026-03-10T11:27:45.051 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-10T11:27:45.051 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm06.local 2026-03-10T11:27:45.051 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T11:27:45.051 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-10T11:27:45.087 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo dnf remove nvme-cli -y 2026-03-10T11:27:45.087 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-10T11:27:45.087 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm09.local 2026-03-10T11:27:45.087 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T11:27:45.087 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-10T11:27:45.268 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: nvme-cli 2026-03-10T11:27:45.268 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:27:45.272 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:27:45.272 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:27:45.273 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:27:45.315 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: nvme-cli 2026-03-10T11:27:45.315 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:27:45.318 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:27:45.318 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:27:45.318 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:27:45.671 INFO:teuthology.orchestra.run.vm06.stdout:Last metadata expiration check: 0:01:29 ago on Tue 10 Mar 2026 11:26:16 AM UTC. 2026-03-10T11:27:45.775 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout: Package Architecture Version Repository Size 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout:Install 6 Packages 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 2.3 M 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout:Installed size: 11 M 2026-03-10T11:27:45.776 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-10T11:27:45.829 INFO:teuthology.orchestra.run.vm09.stdout:Last metadata expiration check: 0:01:25 ago on Tue 10 Mar 2026 11:26:20 AM UTC. 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout: Package Architecture Version Repository Size 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout:Installing: 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout:Installing dependencies: 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout:Install 6 Packages 2026-03-10T11:27:45.956 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:27:45.957 INFO:teuthology.orchestra.run.vm09.stdout:Total download size: 2.3 M 2026-03-10T11:27:45.957 INFO:teuthology.orchestra.run.vm09.stdout:Installed size: 11 M 2026-03-10T11:27:45.957 INFO:teuthology.orchestra.run.vm09.stdout:Downloading Packages: 2026-03-10T11:27:46.382 INFO:teuthology.orchestra.run.vm06.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 312 kB/s | 44 kB 00:00 2026-03-10T11:27:46.388 INFO:teuthology.orchestra.run.vm06.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 491 kB/s | 72 kB 00:00 2026-03-10T11:27:46.461 INFO:teuthology.orchestra.run.vm06.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.0 MB/s | 84 kB 00:00 2026-03-10T11:27:46.469 INFO:teuthology.orchestra.run.vm06.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.8 MB/s | 150 kB 00:00 2026-03-10T11:27:46.533 INFO:teuthology.orchestra.run.vm06.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 3.9 MB/s | 1.2 MB 00:00 2026-03-10T11:27:46.576 INFO:teuthology.orchestra.run.vm09.stdout:(1/6): python3-configshell-1.1.30-1.el9.noarch. 483 kB/s | 72 kB 00:00 2026-03-10T11:27:46.579 INFO:teuthology.orchestra.run.vm09.stdout:(2/6): nvmetcli-0.8-3.el9.noarch.rpm 289 kB/s | 44 kB 00:00 2026-03-10T11:27:46.607 INFO:teuthology.orchestra.run.vm06.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 5.6 MB/s | 837 kB 00:00 2026-03-10T11:27:46.607 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-10T11:27:46.607 INFO:teuthology.orchestra.run.vm06.stdout:Total 2.8 MB/s | 2.3 MB 00:00 2026-03-10T11:27:46.624 INFO:teuthology.orchestra.run.vm09.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.8 MB/s | 84 kB 00:00 2026-03-10T11:27:46.655 INFO:teuthology.orchestra.run.vm09.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.9 MB/s | 150 kB 00:00 2026-03-10T11:27:46.675 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T11:27:46.683 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T11:27:46.683 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T11:27:46.690 INFO:teuthology.orchestra.run.vm09.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 4.4 MB/s | 1.2 MB 00:00 2026-03-10T11:27:46.735 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T11:27:46.735 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T11:27:46.736 INFO:teuthology.orchestra.run.vm09.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 7.3 MB/s | 837 kB 00:00 2026-03-10T11:27:46.736 INFO:teuthology.orchestra.run.vm09.stdout:-------------------------------------------------------------------------------- 2026-03-10T11:27:46.736 INFO:teuthology.orchestra.run.vm09.stdout:Total 3.0 MB/s | 2.3 MB 00:00 2026-03-10T11:27:46.809 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T11:27:46.815 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T11:27:46.815 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T11:27:46.879 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T11:27:46.879 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T11:27:46.879 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T11:27:47.017 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-10T11:27:47.050 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-10T11:27:47.154 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T11:27:47.173 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T11:27:47.235 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-10T11:27:47.347 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T11:27:47.398 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-10T11:27:47.523 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-10T11:27:47.595 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-10T11:27:47.648 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T11:27:47.672 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T11:27:47.681 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T11:27:47.685 INFO:teuthology.orchestra.run.vm09.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-10T11:27:47.882 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-10T11:27:47.891 INFO:teuthology.orchestra.run.vm09.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T11:27:48.007 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T11:27:48.007 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T11:27:48.007 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:27:48.285 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T11:27:48.285 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T11:27:48.285 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:27:48.601 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-10T11:27:48.602 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-10T11:27:48.602 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T11:27:48.602 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T11:27:48.602 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-10T11:27:48.703 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-10T11:27:48.704 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:27:48.704 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-10T11:27:48.704 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T11:27:48.704 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T11:27:48.704 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-10T11:27:48.704 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:27:48.704 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:27:48.794 DEBUG:teuthology.parallel:result is None 2026-03-10T11:27:48.933 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-10T11:27:48.933 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-10T11:27:48.933 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T11:27:48.933 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T11:27:48.933 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-10T11:27:49.117 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-10T11:27:49.117 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:27:49.117 INFO:teuthology.orchestra.run.vm09.stdout:Installed: 2026-03-10T11:27:49.117 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T11:27:49.117 INFO:teuthology.orchestra.run.vm09.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T11:27:49.117 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-10T11:27:49.117 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:27:49.117 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:27:49.187 DEBUG:teuthology.parallel:result is None 2026-03-10T11:27:49.187 INFO:teuthology.run_tasks:Running task install... 2026-03-10T11:27:49.189 DEBUG:teuthology.task.install:project ceph 2026-03-10T11:27:49.189 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'extra_system_packages': {'deb': ['python3-pytest'], 'rpm': ['python3-pytest']}, 'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'}, 'extra_packages': ['cephadm'], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-10T11:27:49.189 DEBUG:teuthology.task.install:config {'extra_system_packages': {'deb': ['python3-pytest', 'python3-xmltodict', 'python3-jmespath'], 'rpm': ['python3-pytest', 'bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-10T11:27:49.189 INFO:teuthology.task.install:Using flavor: default 2026-03-10T11:27:49.191 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-10T11:27:49.191 INFO:teuthology.task.install:extra packages: [] 2026-03-10T11:27:49.192 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-pytest', 'python3-xmltodict', 'python3-jmespath'], 'rpm': ['python3-pytest', 'bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-10T11:27:49.192 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T11:27:49.192 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-pytest', 'python3-xmltodict', 'python3-jmespath'], 'rpm': ['python3-pytest', 'bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-10T11:27:49.192 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T11:27:49.777 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-10T11:27:49.777 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-10T11:27:49.840 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-10T11:27:49.840 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-10T11:27:50.332 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-10T11:27:50.333 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:27:50.333 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-10T11:27:50.334 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-10T11:27:50.334 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:27:50.334 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-10T11:27:50.365 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, python3-pytest, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-10T11:27:50.365 DEBUG:teuthology.orchestra.run.vm06:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-10T11:27:50.375 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, python3-pytest, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-10T11:27:50.376 DEBUG:teuthology.orchestra.run.vm09:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-10T11:27:50.436 DEBUG:teuthology.orchestra.run.vm06:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-10T11:27:50.457 DEBUG:teuthology.orchestra.run.vm09:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-10T11:27:50.522 DEBUG:teuthology.orchestra.run.vm06:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-10T11:27:50.552 INFO:teuthology.orchestra.run.vm06.stdout:check_obsoletes = 1 2026-03-10T11:27:50.553 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-03-10T11:27:50.555 DEBUG:teuthology.orchestra.run.vm09:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-10T11:27:50.585 INFO:teuthology.orchestra.run.vm09.stdout:check_obsoletes = 1 2026-03-10T11:27:50.587 DEBUG:teuthology.orchestra.run.vm09:> sudo yum clean all 2026-03-10T11:27:50.750 INFO:teuthology.orchestra.run.vm06.stdout:41 files removed 2026-03-10T11:27:50.779 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd python3-pytest bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-10T11:27:50.813 INFO:teuthology.orchestra.run.vm09.stdout:41 files removed 2026-03-10T11:27:50.837 DEBUG:teuthology.orchestra.run.vm09:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd python3-pytest bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-10T11:27:52.171 INFO:teuthology.orchestra.run.vm06.stdout:ceph packages for x86_64 71 kB/s | 84 kB 00:01 2026-03-10T11:27:52.270 INFO:teuthology.orchestra.run.vm09.stdout:ceph packages for x86_64 69 kB/s | 84 kB 00:01 2026-03-10T11:27:53.160 INFO:teuthology.orchestra.run.vm06.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-10T11:27:53.238 INFO:teuthology.orchestra.run.vm09.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-10T11:27:54.112 INFO:teuthology.orchestra.run.vm06.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-10T11:27:54.217 INFO:teuthology.orchestra.run.vm09.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-10T11:27:55.084 INFO:teuthology.orchestra.run.vm09.stdout:CentOS Stream 9 - BaseOS 11 MB/s | 8.9 MB 00:00 2026-03-10T11:27:55.275 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - BaseOS 7.8 MB/s | 8.9 MB 00:01 2026-03-10T11:27:57.406 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - AppStream 19 MB/s | 27 MB 00:01 2026-03-10T11:28:02.410 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - CRB 3.4 MB/s | 8.0 MB 00:02 2026-03-10T11:28:03.544 INFO:teuthology.orchestra.run.vm09.stdout:CentOS Stream 9 - AppStream 3.6 MB/s | 27 MB 00:07 2026-03-10T11:28:03.833 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - Extras packages 35 kB/s | 20 kB 00:00 2026-03-10T11:28:04.307 INFO:teuthology.orchestra.run.vm06.stdout:Extra Packages for Enterprise Linux 52 MB/s | 20 MB 00:00 2026-03-10T11:28:06.606 INFO:teuthology.orchestra.run.vm09.stdout:CentOS Stream 9 - CRB 22 MB/s | 8.0 MB 00:00 2026-03-10T11:28:08.085 INFO:teuthology.orchestra.run.vm09.stdout:CentOS Stream 9 - Extras packages 34 kB/s | 20 kB 00:00 2026-03-10T11:28:08.546 INFO:teuthology.orchestra.run.vm09.stdout:Extra Packages for Enterprise Linux 54 MB/s | 20 MB 00:00 2026-03-10T11:28:09.314 INFO:teuthology.orchestra.run.vm06.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-10T11:28:10.764 INFO:teuthology.orchestra.run.vm06.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T11:28:10.765 INFO:teuthology.orchestra.run.vm06.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T11:28:10.769 INFO:teuthology.orchestra.run.vm06.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-10T11:28:10.770 INFO:teuthology.orchestra.run.vm06.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-10T11:28:10.797 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:28:10.801 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-10T11:28:10.801 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T11:28:10.801 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-10T11:28:10.801 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytest noarch 6.2.2-7.el9 appstream 519 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout:Upgrading: 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-10T11:28:10.802 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-iniconfig noarch 1.1.1-7.el9 appstream 17 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-pluggy noarch 0.13.1-7.el9 appstream 41 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-py noarch 1.10.0-6.el9 appstream 477 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-10T11:28:10.803 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout:Installing weak dependencies: 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout:Install 138 Packages 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout:Upgrade 2 Packages 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 211 M 2026-03-10T11:28:10.804 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-10T11:28:12.587 INFO:teuthology.orchestra.run.vm06.stdout:(1/140): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-10T11:28:13.416 INFO:teuthology.orchestra.run.vm06.stdout:(2/140): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.4 MB/s | 1.2 MB 00:00 2026-03-10T11:28:13.536 INFO:teuthology.orchestra.run.vm06.stdout:(3/140): ceph-immutable-object-cache-19.2.3-678 1.2 MB/s | 145 kB 00:00 2026-03-10T11:28:13.615 INFO:teuthology.orchestra.run.vm06.stdout:(4/140): ceph-base-19.2.3-678.ge911bdeb.el9.x86 3.7 MB/s | 5.5 MB 00:01 2026-03-10T11:28:13.741 INFO:teuthology.orchestra.run.vm06.stdout:(5/140): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 8.5 MB/s | 1.1 MB 00:00 2026-03-10T11:28:13.805 INFO:teuthology.orchestra.run.vm06.stdout:(6/140): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 9.0 MB/s | 2.4 MB 00:00 2026-03-10T11:28:13.815 INFO:teuthology.orchestra.run.vm09.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-10T11:28:14.110 INFO:teuthology.orchestra.run.vm06.stdout:(7/140): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 13 MB/s | 4.7 MB 00:00 2026-03-10T11:28:14.680 INFO:teuthology.orchestra.run.vm06.stdout:(8/140): ceph-common-19.2.3-678.ge911bdeb.el9.x 8.5 MB/s | 22 MB 00:02 2026-03-10T11:28:14.798 INFO:teuthology.orchestra.run.vm06.stdout:(9/140): ceph-selinux-19.2.3-678.ge911bdeb.el9. 213 kB/s | 25 kB 00:00 2026-03-10T11:28:14.836 INFO:teuthology.orchestra.run.vm06.stdout:(10/140): ceph-radosgw-19.2.3-678.ge911bdeb.el9 15 MB/s | 11 MB 00:00 2026-03-10T11:28:14.951 INFO:teuthology.orchestra.run.vm06.stdout:(11/140): libcephfs-devel-19.2.3-678.ge911bdeb. 292 kB/s | 34 kB 00:00 2026-03-10T11:28:15.043 INFO:teuthology.orchestra.run.vm06.stdout:(12/140): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 14 MB/s | 17 MB 00:01 2026-03-10T11:28:15.077 INFO:teuthology.orchestra.run.vm06.stdout:(13/140): libcephfs2-19.2.3-678.ge911bdeb.el9.x 7.8 MB/s | 1.0 MB 00:00 2026-03-10T11:28:15.168 INFO:teuthology.orchestra.run.vm06.stdout:(14/140): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-10T11:28:15.193 INFO:teuthology.orchestra.run.vm06.stdout:(15/140): librados-devel-19.2.3-678.ge911bdeb.e 1.1 MB/s | 127 kB 00:00 2026-03-10T11:28:15.225 INFO:teuthology.orchestra.run.vm09.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T11:28:15.226 INFO:teuthology.orchestra.run.vm09.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T11:28:15.233 INFO:teuthology.orchestra.run.vm09.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-10T11:28:15.233 INFO:teuthology.orchestra.run.vm09.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-10T11:28:15.262 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout:====================================================================================== 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout:====================================================================================== 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout:Installing: 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: python3-pytest noarch 6.2.2-7.el9 appstream 519 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout:Upgrading: 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout:Installing dependencies: 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-10T11:28:15.267 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-iniconfig noarch 1.1.1-7.el9 appstream 17 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-10T11:28:15.268 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-pluggy noarch 0.13.1-7.el9 appstream 41 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-py noarch 1.10.0-6.el9 appstream 477 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout:Installing weak dependencies: 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout:====================================================================================== 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout:Install 138 Packages 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout:Upgrade 2 Packages 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout:Total download size: 211 M 2026-03-10T11:28:15.269 INFO:teuthology.orchestra.run.vm09.stdout:Downloading Packages: 2026-03-10T11:28:15.294 INFO:teuthology.orchestra.run.vm06.stdout:(16/140): libradosstriper1-19.2.3-678.ge911bdeb 3.9 MB/s | 503 kB 00:00 2026-03-10T11:28:15.413 INFO:teuthology.orchestra.run.vm06.stdout:(17/140): python3-ceph-argparse-19.2.3-678.ge91 379 kB/s | 45 kB 00:00 2026-03-10T11:28:15.535 INFO:teuthology.orchestra.run.vm06.stdout:(18/140): python3-ceph-common-19.2.3-678.ge911b 1.1 MB/s | 142 kB 00:00 2026-03-10T11:28:15.571 INFO:teuthology.orchestra.run.vm06.stdout:(19/140): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 14 MB/s | 5.4 MB 00:00 2026-03-10T11:28:15.656 INFO:teuthology.orchestra.run.vm06.stdout:(20/140): python3-cephfs-19.2.3-678.ge911bdeb.e 1.3 MB/s | 165 kB 00:00 2026-03-10T11:28:15.689 INFO:teuthology.orchestra.run.vm06.stdout:(21/140): python3-rados-19.2.3-678.ge911bdeb.el 2.7 MB/s | 323 kB 00:00 2026-03-10T11:28:15.777 INFO:teuthology.orchestra.run.vm06.stdout:(22/140): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 303 kB 00:00 2026-03-10T11:28:15.804 INFO:teuthology.orchestra.run.vm06.stdout:(23/140): python3-rgw-19.2.3-678.ge911bdeb.el9. 864 kB/s | 100 kB 00:00 2026-03-10T11:28:15.985 INFO:teuthology.orchestra.run.vm06.stdout:(24/140): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 410 kB/s | 85 kB 00:00 2026-03-10T11:28:16.136 INFO:teuthology.orchestra.run.vm06.stdout:(25/140): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 9.4 MB/s | 3.1 MB 00:00 2026-03-10T11:28:16.142 INFO:teuthology.orchestra.run.vm06.stdout:(26/140): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.1 MB/s | 171 kB 00:00 2026-03-10T11:28:16.262 INFO:teuthology.orchestra.run.vm06.stdout:(27/140): ceph-grafana-dashboards-19.2.3-678.ge 247 kB/s | 31 kB 00:00 2026-03-10T11:28:16.269 INFO:teuthology.orchestra.run.vm06.stdout:(28/140): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-10T11:28:16.534 INFO:teuthology.orchestra.run.vm06.stdout:(29/140): ceph-mgr-dashboard-19.2.3-678.ge911bd 14 MB/s | 3.8 MB 00:00 2026-03-10T11:28:16.652 INFO:teuthology.orchestra.run.vm06.stdout:(30/140): ceph-mgr-modules-core-19.2.3-678.ge91 2.1 MB/s | 253 kB 00:00 2026-03-10T11:28:16.768 INFO:teuthology.orchestra.run.vm06.stdout:(31/140): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 426 kB/s | 49 kB 00:00 2026-03-10T11:28:16.815 INFO:teuthology.orchestra.run.vm06.stdout:(32/140): ceph-mgr-diskprediction-local-19.2.3- 14 MB/s | 7.4 MB 00:00 2026-03-10T11:28:16.883 INFO:teuthology.orchestra.run.vm06.stdout:(33/140): ceph-prometheus-alerts-19.2.3-678.ge9 146 kB/s | 17 kB 00:00 2026-03-10T11:28:16.911 INFO:teuthology.orchestra.run.vm09.stdout:(1/140): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 13 kB/s | 6.5 kB 00:00 2026-03-10T11:28:16.939 INFO:teuthology.orchestra.run.vm06.stdout:(34/140): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 299 kB 00:00 2026-03-10T11:28:17.007 INFO:teuthology.orchestra.run.vm06.stdout:(35/140): cephadm-19.2.3-678.ge911bdeb.el9.noar 6.1 MB/s | 769 kB 00:00 2026-03-10T11:28:17.151 INFO:teuthology.orchestra.run.vm06.stdout:(36/140): ledmon-libs-1.1.0-3.el9.x86_64.rpm 281 kB/s | 40 kB 00:00 2026-03-10T11:28:17.248 INFO:teuthology.orchestra.run.vm06.stdout:(37/140): libconfig-1.7.2-9.el9.x86_64.rpm 748 kB/s | 72 kB 00:00 2026-03-10T11:28:17.284 INFO:teuthology.orchestra.run.vm06.stdout:(38/140): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.0 MB/s | 351 kB 00:00 2026-03-10T11:28:17.349 INFO:teuthology.orchestra.run.vm06.stdout:(39/140): libquadmath-11.5.0-14.el9.x86_64.rpm 2.8 MB/s | 184 kB 00:00 2026-03-10T11:28:17.402 INFO:teuthology.orchestra.run.vm06.stdout:(40/140): mailcap-2.1.49-5.el9.noarch.rpm 629 kB/s | 33 kB 00:00 2026-03-10T11:28:17.425 INFO:teuthology.orchestra.run.vm06.stdout:(41/140): libgfortran-11.5.0-14.el9.x86_64.rpm 4.4 MB/s | 794 kB 00:00 2026-03-10T11:28:17.461 INFO:teuthology.orchestra.run.vm06.stdout:(42/140): pciutils-3.7.0-7.el9.x86_64.rpm 1.6 MB/s | 93 kB 00:00 2026-03-10T11:28:17.490 INFO:teuthology.orchestra.run.vm06.stdout:(43/140): python3-cffi-1.14.5-5.el9.x86_64.rpm 3.8 MB/s | 253 kB 00:00 2026-03-10T11:28:17.553 INFO:teuthology.orchestra.run.vm06.stdout:(44/140): python3-ply-3.11-14.el9.noarch.rpm 1.7 MB/s | 106 kB 00:00 2026-03-10T11:28:17.607 INFO:teuthology.orchestra.run.vm06.stdout:(45/140): python3-cryptography-36.0.1-5.el9.x86 8.5 MB/s | 1.2 MB 00:00 2026-03-10T11:28:17.616 INFO:teuthology.orchestra.run.vm06.stdout:(46/140): python3-pycparser-2.20-6.el9.noarch.r 2.1 MB/s | 135 kB 00:00 2026-03-10T11:28:17.665 INFO:teuthology.orchestra.run.vm06.stdout:(47/140): python3-requests-2.25.1-10.el9.noarch 2.1 MB/s | 126 kB 00:00 2026-03-10T11:28:17.677 INFO:teuthology.orchestra.run.vm06.stdout:(48/140): python3-urllib3-1.26.5-7.el9.noarch.r 3.5 MB/s | 218 kB 00:00 2026-03-10T11:28:17.749 INFO:teuthology.orchestra.run.vm06.stdout:(49/140): unzip-6.0-59.el9.x86_64.rpm 2.1 MB/s | 182 kB 00:00 2026-03-10T11:28:17.763 INFO:teuthology.orchestra.run.vm09.stdout:(2/140): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.4 MB/s | 1.2 MB 00:00 2026-03-10T11:28:17.858 INFO:teuthology.orchestra.run.vm06.stdout:(50/140): boost-program-options-1.75.0-13.el9.x 956 kB/s | 104 kB 00:00 2026-03-10T11:28:17.886 INFO:teuthology.orchestra.run.vm09.stdout:(3/140): ceph-immutable-object-cache-19.2.3-678 1.2 MB/s | 145 kB 00:00 2026-03-10T11:28:18.027 INFO:teuthology.orchestra.run.vm06.stdout:(51/140): ceph-test-19.2.3-678.ge911bdeb.el9.x8 15 MB/s | 50 MB 00:03 2026-03-10T11:28:18.028 INFO:teuthology.orchestra.run.vm06.stdout:(52/140): zip-3.0-35.el9.x86_64.rpm 756 kB/s | 266 kB 00:00 2026-03-10T11:28:18.029 INFO:teuthology.orchestra.run.vm06.stdout:(53/140): flexiblas-3.0.4-9.el9.x86_64.rpm 173 kB/s | 30 kB 00:00 2026-03-10T11:28:18.037 INFO:teuthology.orchestra.run.vm09.stdout:(4/140): ceph-base-19.2.3-678.ge911bdeb.el9.x86 3.4 MB/s | 5.5 MB 00:01 2026-03-10T11:28:18.078 INFO:teuthology.orchestra.run.vm06.stdout:(54/140): libnbd-1.20.3-4.el9.x86_64.rpm 3.3 MB/s | 164 kB 00:00 2026-03-10T11:28:18.123 INFO:teuthology.orchestra.run.vm06.stdout:(55/140): libpmemobj-1.12.1-1.el9.x86_64.rpm 3.5 MB/s | 160 kB 00:00 2026-03-10T11:28:18.135 INFO:teuthology.orchestra.run.vm06.stdout:(56/140): flexiblas-openblas-openmp-3.0.4-9.el9 140 kB/s | 15 kB 00:00 2026-03-10T11:28:18.144 INFO:teuthology.orchestra.run.vm09.stdout:(5/140): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 9.4 MB/s | 2.4 MB 00:00 2026-03-10T11:28:18.147 INFO:teuthology.orchestra.run.vm06.stdout:(57/140): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.8 MB/s | 45 kB 00:00 2026-03-10T11:28:18.163 INFO:teuthology.orchestra.run.vm09.stdout:(6/140): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 8.5 MB/s | 1.1 MB 00:00 2026-03-10T11:28:18.170 INFO:teuthology.orchestra.run.vm06.stdout:(58/140): libstoragemgmt-1.10.1-1.el9.x86_64.rp 11 MB/s | 246 kB 00:00 2026-03-10T11:28:18.208 INFO:teuthology.orchestra.run.vm06.stdout:(59/140): libxslt-1.1.34-12.el9.x86_64.rpm 6.1 MB/s | 233 kB 00:00 2026-03-10T11:28:18.246 INFO:teuthology.orchestra.run.vm06.stdout:(60/140): librdkafka-1.6.1-102.el9.x86_64.rpm 5.9 MB/s | 662 kB 00:00 2026-03-10T11:28:18.272 INFO:teuthology.orchestra.run.vm06.stdout:(61/140): flexiblas-netlib-3.0.4-9.el9.x86_64.r 12 MB/s | 3.0 MB 00:00 2026-03-10T11:28:18.274 INFO:teuthology.orchestra.run.vm06.stdout:(62/140): lua-5.4.4-4.el9.x86_64.rpm 6.5 MB/s | 188 kB 00:00 2026-03-10T11:28:18.275 INFO:teuthology.orchestra.run.vm06.stdout:(63/140): lttng-ust-2.12.0-6.el9.x86_64.rpm 4.2 MB/s | 292 kB 00:00 2026-03-10T11:28:18.305 INFO:teuthology.orchestra.run.vm06.stdout:(64/140): openblas-0.3.29-1.el9.x86_64.rpm 1.3 MB/s | 42 kB 00:00 2026-03-10T11:28:18.459 INFO:teuthology.orchestra.run.vm06.stdout:(65/140): openblas-openmp-0.3.29-1.el9.x86_64.r 29 MB/s | 5.3 MB 00:00 2026-03-10T11:28:18.513 INFO:teuthology.orchestra.run.vm06.stdout:(66/140): python3-devel-3.9.25-3.el9.x86_64.rpm 4.3 MB/s | 244 kB 00:00 2026-03-10T11:28:18.533 INFO:teuthology.orchestra.run.vm06.stdout:(67/140): protobuf-3.14.0-17.el9.x86_64.rpm 3.9 MB/s | 1.0 MB 00:00 2026-03-10T11:28:18.549 INFO:teuthology.orchestra.run.vm06.stdout:(68/140): python3-babel-2.9.1-2.el9.noarch.rpm 24 MB/s | 6.0 MB 00:00 2026-03-10T11:28:18.550 INFO:teuthology.orchestra.run.vm06.stdout:(69/140): python3-iniconfig-1.1.1-7.el9.noarch. 478 kB/s | 17 kB 00:00 2026-03-10T11:28:18.566 INFO:teuthology.orchestra.run.vm06.stdout:(70/140): python3-jinja2-2.11.3-8.el9.noarch.rp 7.3 MB/s | 249 kB 00:00 2026-03-10T11:28:18.575 INFO:teuthology.orchestra.run.vm06.stdout:(71/140): python3-libstoragemgmt-1.10.1-1.el9.x 7.1 MB/s | 177 kB 00:00 2026-03-10T11:28:18.581 INFO:teuthology.orchestra.run.vm06.stdout:(72/140): python3-jmespath-1.0.1-1.el9.noarch.r 1.4 MB/s | 48 kB 00:00 2026-03-10T11:28:18.589 INFO:teuthology.orchestra.run.vm06.stdout:(73/140): python3-mako-1.1.4-6.el9.noarch.rpm 7.2 MB/s | 172 kB 00:00 2026-03-10T11:28:18.598 INFO:teuthology.orchestra.run.vm06.stdout:(74/140): python3-markupsafe-1.1.1-12.el9.x86_6 1.5 MB/s | 35 kB 00:00 2026-03-10T11:28:18.624 INFO:teuthology.orchestra.run.vm06.stdout:(75/140): python3-packaging-20.9-5.el9.noarch.r 2.8 MB/s | 77 kB 00:00 2026-03-10T11:28:18.659 INFO:teuthology.orchestra.run.vm06.stdout:(76/140): python3-numpy-f2py-1.23.5-2.el9.x86_6 6.2 MB/s | 442 kB 00:00 2026-03-10T11:28:18.666 INFO:teuthology.orchestra.run.vm06.stdout:(77/140): python3-pluggy-0.13.1-7.el9.noarch.rp 990 kB/s | 41 kB 00:00 2026-03-10T11:28:18.666 INFO:teuthology.orchestra.run.vm09.stdout:(7/140): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 9.1 MB/s | 4.7 MB 00:00 2026-03-10T11:28:18.692 INFO:teuthology.orchestra.run.vm06.stdout:(78/140): python3-protobuf-3.14.0-17.el9.noarch 8.1 MB/s | 267 kB 00:00 2026-03-10T11:28:18.700 INFO:teuthology.orchestra.run.vm06.stdout:(79/140): python3-py-1.10.0-6.el9.noarch.rpm 14 MB/s | 477 kB 00:00 2026-03-10T11:28:18.726 INFO:teuthology.orchestra.run.vm06.stdout:(80/140): python3-pyasn1-0.4.8-7.el9.noarch.rpm 4.6 MB/s | 157 kB 00:00 2026-03-10T11:28:18.744 INFO:teuthology.orchestra.run.vm06.stdout:(81/140): python3-pyasn1-modules-0.4.8-7.el9.no 6.3 MB/s | 277 kB 00:00 2026-03-10T11:28:18.763 INFO:teuthology.orchestra.run.vm06.stdout:(82/140): python3-requests-oauthlib-1.3.0-12.el 2.8 MB/s | 54 kB 00:00 2026-03-10T11:28:18.777 INFO:teuthology.orchestra.run.vm06.stdout:(83/140): python3-pytest-6.2.2-7.el9.noarch.rpm 10 MB/s | 519 kB 00:00 2026-03-10T11:28:18.797 INFO:teuthology.orchestra.run.vm06.stdout:(84/140): python3-numpy-1.23.5-2.el9.x86_64.rpm 28 MB/s | 6.1 MB 00:00 2026-03-10T11:28:18.814 INFO:teuthology.orchestra.run.vm06.stdout:(85/140): python3-toml-0.10.2-6.el9.noarch.rpm 1.1 MB/s | 42 kB 00:00 2026-03-10T11:28:18.830 INFO:teuthology.orchestra.run.vm06.stdout:(86/140): qatlib-25.08.0-2.el9.x86_64.rpm 7.3 MB/s | 240 kB 00:00 2026-03-10T11:28:18.843 INFO:teuthology.orchestra.run.vm06.stdout:(87/140): qatlib-service-25.08.0-2.el9.x86_64.r 1.3 MB/s | 37 kB 00:00 2026-03-10T11:28:18.860 INFO:teuthology.orchestra.run.vm06.stdout:(88/140): qatzip-libs-1.3.1-1.el9.x86_64.rpm 2.1 MB/s | 66 kB 00:00 2026-03-10T11:28:18.890 INFO:teuthology.orchestra.run.vm06.stdout:(89/140): socat-1.7.4.1-8.el9.x86_64.rpm 6.4 MB/s | 303 kB 00:00 2026-03-10T11:28:18.900 INFO:teuthology.orchestra.run.vm06.stdout:(90/140): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.6 MB/s | 64 kB 00:00 2026-03-10T11:28:18.928 INFO:teuthology.orchestra.run.vm06.stdout:(91/140): lua-devel-5.4.4-4.el9.x86_64.rpm 595 kB/s | 22 kB 00:00 2026-03-10T11:28:18.946 INFO:teuthology.orchestra.run.vm06.stdout:(92/140): abseil-cpp-20211102.0-4.el9.x86_64.rp 29 MB/s | 551 kB 00:00 2026-03-10T11:28:18.957 INFO:teuthology.orchestra.run.vm06.stdout:(93/140): gperftools-libs-2.9.1-3.el9.x86_64.rp 29 MB/s | 308 kB 00:00 2026-03-10T11:28:18.960 INFO:teuthology.orchestra.run.vm06.stdout:(94/140): grpc-data-1.46.7-10.el9.noarch.rpm 6.8 MB/s | 19 kB 00:00 2026-03-10T11:28:19.018 INFO:teuthology.orchestra.run.vm06.stdout:(95/140): protobuf-compiler-3.14.0-17.el9.x86_6 7.2 MB/s | 862 kB 00:00 2026-03-10T11:28:19.025 INFO:teuthology.orchestra.run.vm06.stdout:(96/140): libarrow-doc-9.0.0-15.el9.noarch.rpm 3.9 MB/s | 25 kB 00:00 2026-03-10T11:28:19.034 INFO:teuthology.orchestra.run.vm06.stdout:(97/140): liboath-2.6.12-1.el9.x86_64.rpm 5.4 MB/s | 49 kB 00:00 2026-03-10T11:28:19.049 INFO:teuthology.orchestra.run.vm06.stdout:(98/140): libunwind-1.6.2-1.el9.x86_64.rpm 4.5 MB/s | 67 kB 00:00 2026-03-10T11:28:19.058 INFO:teuthology.orchestra.run.vm06.stdout:(99/140): luarocks-3.9.2-5.el9.noarch.rpm 16 MB/s | 151 kB 00:00 2026-03-10T11:28:19.080 INFO:teuthology.orchestra.run.vm06.stdout:(100/140): libarrow-9.0.0-15.el9.x86_64.rpm 37 MB/s | 4.4 MB 00:00 2026-03-10T11:28:19.093 INFO:teuthology.orchestra.run.vm06.stdout:(101/140): python3-asyncssh-2.13.2-5.el9.noarch 43 MB/s | 548 kB 00:00 2026-03-10T11:28:19.095 INFO:teuthology.orchestra.run.vm06.stdout:(102/140): python3-autocommand-2.2.2-8.el9.noar 12 MB/s | 29 kB 00:00 2026-03-10T11:28:19.101 INFO:teuthology.orchestra.run.vm06.stdout:(103/140): python3-backports-tarfile-1.2.0-1.el 11 MB/s | 60 kB 00:00 2026-03-10T11:28:19.106 INFO:teuthology.orchestra.run.vm06.stdout:(104/140): parquet-libs-9.0.0-15.el9.x86_64.rpm 17 MB/s | 838 kB 00:00 2026-03-10T11:28:19.106 INFO:teuthology.orchestra.run.vm06.stdout:(105/140): python3-bcrypt-3.2.2-1.el9.x86_64.rp 8.1 MB/s | 43 kB 00:00 2026-03-10T11:28:19.108 INFO:teuthology.orchestra.run.vm06.stdout:(106/140): python3-cachetools-4.2.4-1.el9.noarc 14 MB/s | 32 kB 00:00 2026-03-10T11:28:19.109 INFO:teuthology.orchestra.run.vm06.stdout:(107/140): python3-certifi-2023.05.07-4.el9.noa 6.4 MB/s | 14 kB 00:00 2026-03-10T11:28:19.118 INFO:teuthology.orchestra.run.vm06.stdout:(108/140): python3-cheroot-10.0.1-4.el9.noarch. 17 MB/s | 173 kB 00:00 2026-03-10T11:28:19.120 INFO:teuthology.orchestra.run.vm06.stdout:(109/140): python3-cherrypy-18.6.1-2.el9.noarch 31 MB/s | 358 kB 00:00 2026-03-10T11:28:19.127 INFO:teuthology.orchestra.run.vm06.stdout:(110/140): python3-google-auth-2.45.0-1.el9.noa 29 MB/s | 254 kB 00:00 2026-03-10T11:28:19.134 INFO:teuthology.orchestra.run.vm06.stdout:(111/140): python3-grpcio-tools-1.46.7-10.el9.x 22 MB/s | 144 kB 00:00 2026-03-10T11:28:19.137 INFO:teuthology.orchestra.run.vm06.stdout:(112/140): python3-jaraco-8.2.1-3.el9.noarch.rp 3.3 MB/s | 11 kB 00:00 2026-03-10T11:28:19.140 INFO:teuthology.orchestra.run.vm06.stdout:(113/140): python3-jaraco-classes-3.2.1-5.el9.n 5.9 MB/s | 18 kB 00:00 2026-03-10T11:28:19.144 INFO:teuthology.orchestra.run.vm06.stdout:(114/140): python3-jaraco-collections-3.0.0-8.e 6.8 MB/s | 23 kB 00:00 2026-03-10T11:28:19.147 INFO:teuthology.orchestra.run.vm06.stdout:(115/140): python3-jaraco-context-6.0.1-3.el9.n 7.1 MB/s | 20 kB 00:00 2026-03-10T11:28:19.150 INFO:teuthology.orchestra.run.vm06.stdout:(116/140): python3-jaraco-functools-3.5.0-2.el9 5.5 MB/s | 19 kB 00:00 2026-03-10T11:28:19.155 INFO:teuthology.orchestra.run.vm06.stdout:(117/140): python3-jaraco-text-4.0.0-2.el9.noar 5.3 MB/s | 26 kB 00:00 2026-03-10T11:28:19.172 INFO:teuthology.orchestra.run.vm06.stdout:(118/140): python3-grpcio-1.46.7-10.el9.x86_64. 40 MB/s | 2.0 MB 00:00 2026-03-10T11:28:19.175 INFO:teuthology.orchestra.run.vm06.stdout:(119/140): python3-logutils-0.3.5-21.el9.noarch 14 MB/s | 46 kB 00:00 2026-03-10T11:28:19.179 INFO:teuthology.orchestra.run.vm06.stdout:(120/140): python3-more-itertools-8.12.0-2.el9. 21 MB/s | 79 kB 00:00 2026-03-10T11:28:19.184 INFO:teuthology.orchestra.run.vm06.stdout:(121/140): python3-kubernetes-26.1.0-3.el9.noar 36 MB/s | 1.0 MB 00:00 2026-03-10T11:28:19.185 INFO:teuthology.orchestra.run.vm06.stdout:(122/140): python3-natsort-7.1.1-5.el9.noarch.r 9.8 MB/s | 58 kB 00:00 2026-03-10T11:28:19.189 INFO:teuthology.orchestra.run.vm06.stdout:(123/140): python3-portend-3.1.0-2.el9.noarch.r 3.8 MB/s | 16 kB 00:00 2026-03-10T11:28:19.195 INFO:teuthology.orchestra.run.vm06.stdout:(124/140): python3-pecan-1.4.2-3.el9.noarch.rpm 25 MB/s | 272 kB 00:00 2026-03-10T11:28:19.197 INFO:teuthology.orchestra.run.vm06.stdout:(125/140): python3-pyOpenSSL-21.0.0-1.el9.noarc 12 MB/s | 90 kB 00:00 2026-03-10T11:28:19.199 INFO:teuthology.orchestra.run.vm06.stdout:(126/140): python3-repoze-lru-0.7-16.el9.noarch 7.9 MB/s | 31 kB 00:00 2026-03-10T11:28:19.202 INFO:teuthology.orchestra.run.vm06.stdout:(127/140): python3-routes-2.5.1-5.el9.noarch.rp 40 MB/s | 188 kB 00:00 2026-03-10T11:28:19.202 INFO:teuthology.orchestra.run.vm06.stdout:(128/140): python3-rsa-4.9-2.el9.noarch.rpm 16 MB/s | 59 kB 00:00 2026-03-10T11:28:19.204 INFO:teuthology.orchestra.run.vm06.stdout:(129/140): python3-tempora-5.0.0-2.el9.noarch.r 14 MB/s | 36 kB 00:00 2026-03-10T11:28:19.206 INFO:teuthology.orchestra.run.vm06.stdout:(130/140): python3-typing-extensions-4.15.0-1.e 27 MB/s | 86 kB 00:00 2026-03-10T11:28:19.214 INFO:teuthology.orchestra.run.vm06.stdout:(131/140): python3-websocket-client-1.2.3-2.el9 11 MB/s | 90 kB 00:00 2026-03-10T11:28:19.216 INFO:teuthology.orchestra.run.vm06.stdout:(132/140): python3-webob-1.8.8-2.el9.noarch.rpm 19 MB/s | 230 kB 00:00 2026-03-10T11:28:19.220 INFO:teuthology.orchestra.run.vm06.stdout:(133/140): python3-xmltodict-0.12.0-15.el9.noar 6.4 MB/s | 22 kB 00:00 2026-03-10T11:28:19.223 INFO:teuthology.orchestra.run.vm06.stdout:(134/140): python3-werkzeug-2.0.3-3.el9.1.noarc 47 MB/s | 427 kB 00:00 2026-03-10T11:28:19.224 INFO:teuthology.orchestra.run.vm06.stdout:(135/140): python3-zc-lockfile-2.0-10.el9.noarc 5.0 MB/s | 20 kB 00:00 2026-03-10T11:28:19.228 INFO:teuthology.orchestra.run.vm06.stdout:(136/140): re2-20211101-20.el9.x86_64.rpm 36 MB/s | 191 kB 00:00 2026-03-10T11:28:19.291 INFO:teuthology.orchestra.run.vm06.stdout:(137/140): python3-scipy-1.9.3-2.el9.x86_64.rpm 37 MB/s | 19 MB 00:00 2026-03-10T11:28:19.300 INFO:teuthology.orchestra.run.vm06.stdout:(138/140): thrift-0.15.0-4.el9.x86_64.rpm 21 MB/s | 1.6 MB 00:00 2026-03-10T11:28:20.021 INFO:teuthology.orchestra.run.vm09.stdout:(8/140): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 7.9 MB/s | 11 MB 00:01 2026-03-10T11:28:20.144 INFO:teuthology.orchestra.run.vm09.stdout:(9/140): ceph-selinux-19.2.3-678.ge911bdeb.el9. 205 kB/s | 25 kB 00:00 2026-03-10T11:28:20.189 INFO:teuthology.orchestra.run.vm06.stdout:(139/140): librados2-19.2.3-678.ge911bdeb.el9.x 3.6 MB/s | 3.4 MB 00:00 2026-03-10T11:28:20.217 INFO:teuthology.orchestra.run.vm06.stdout:(140/140): librbd1-19.2.3-678.ge911bdeb.el9.x86 3.4 MB/s | 3.2 MB 00:00 2026-03-10T11:28:20.221 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-10T11:28:20.221 INFO:teuthology.orchestra.run.vm06.stdout:Total 22 MB/s | 211 MB 00:09 2026-03-10T11:28:20.769 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T11:28:20.818 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T11:28:20.818 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T11:28:21.037 INFO:teuthology.orchestra.run.vm09.stdout:(10/140): ceph-common-19.2.3-678.ge911bdeb.el9. 4.7 MB/s | 22 MB 00:04 2026-03-10T11:28:21.087 INFO:teuthology.orchestra.run.vm09.stdout:(11/140): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 5.8 MB/s | 17 MB 00:02 2026-03-10T11:28:21.157 INFO:teuthology.orchestra.run.vm09.stdout:(12/140): libcephfs-devel-19.2.3-678.ge911bdeb. 280 kB/s | 34 kB 00:00 2026-03-10T11:28:21.279 INFO:teuthology.orchestra.run.vm09.stdout:(13/140): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-10T11:28:21.320 INFO:teuthology.orchestra.run.vm09.stdout:(14/140): libcephfs2-19.2.3-678.ge911bdeb.el9.x 4.2 MB/s | 1.0 MB 00:00 2026-03-10T11:28:21.399 INFO:teuthology.orchestra.run.vm09.stdout:(15/140): librados-devel-19.2.3-678.ge911bdeb.e 1.0 MB/s | 127 kB 00:00 2026-03-10T11:28:21.440 INFO:teuthology.orchestra.run.vm09.stdout:(16/140): libradosstriper1-19.2.3-678.ge911bdeb 4.1 MB/s | 503 kB 00:00 2026-03-10T11:28:21.555 INFO:teuthology.orchestra.run.vm09.stdout:(17/140): python3-ceph-argparse-19.2.3-678.ge91 393 kB/s | 45 kB 00:00 2026-03-10T11:28:21.637 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T11:28:21.637 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T11:28:21.671 INFO:teuthology.orchestra.run.vm09.stdout:(18/140): python3-ceph-common-19.2.3-678.ge911b 1.2 MB/s | 142 kB 00:00 2026-03-10T11:28:21.786 INFO:teuthology.orchestra.run.vm09.stdout:(19/140): python3-cephfs-19.2.3-678.ge911bdeb.e 1.4 MB/s | 165 kB 00:00 2026-03-10T11:28:21.904 INFO:teuthology.orchestra.run.vm09.stdout:(20/140): python3-rados-19.2.3-678.ge911bdeb.el 2.7 MB/s | 323 kB 00:00 2026-03-10T11:28:22.021 INFO:teuthology.orchestra.run.vm09.stdout:(21/140): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.5 MB/s | 303 kB 00:00 2026-03-10T11:28:22.137 INFO:teuthology.orchestra.run.vm09.stdout:(22/140): python3-rgw-19.2.3-678.ge911bdeb.el9. 866 kB/s | 100 kB 00:00 2026-03-10T11:28:22.252 INFO:teuthology.orchestra.run.vm09.stdout:(23/140): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 739 kB/s | 85 kB 00:00 2026-03-10T11:28:22.369 INFO:teuthology.orchestra.run.vm09.stdout:(24/140): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 5.6 MB/s | 5.4 MB 00:00 2026-03-10T11:28:22.491 INFO:teuthology.orchestra.run.vm09.stdout:(25/140): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.4 MB/s | 171 kB 00:00 2026-03-10T11:28:22.527 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T11:28:22.542 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/142 2026-03-10T11:28:22.554 INFO:teuthology.orchestra.run.vm06.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/142 2026-03-10T11:28:22.611 INFO:teuthology.orchestra.run.vm09.stdout:(26/140): ceph-grafana-dashboards-19.2.3-678.ge 259 kB/s | 31 kB 00:00 2026-03-10T11:28:22.720 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/142 2026-03-10T11:28:22.723 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/142 2026-03-10T11:28:22.733 INFO:teuthology.orchestra.run.vm09.stdout:(27/140): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-10T11:28:22.782 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/142 2026-03-10T11:28:22.784 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/142 2026-03-10T11:28:22.816 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/142 2026-03-10T11:28:22.826 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/142 2026-03-10T11:28:22.830 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/142 2026-03-10T11:28:22.834 INFO:teuthology.orchestra.run.vm09.stdout:(28/140): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 5.4 MB/s | 3.1 MB 00:00 2026-03-10T11:28:22.837 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/142 2026-03-10T11:28:22.849 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/142 2026-03-10T11:28:22.856 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-packaging-20.9-5.el9.noarch 10/142 2026-03-10T11:28:22.866 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 11/142 2026-03-10T11:28:22.868 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 12/142 2026-03-10T11:28:22.906 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 12/142 2026-03-10T11:28:22.907 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 13/142 2026-03-10T11:28:22.925 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 13/142 2026-03-10T11:28:22.959 INFO:teuthology.orchestra.run.vm06.stdout: Installing : re2-1:20211101-20.el9.x86_64 14/142 2026-03-10T11:28:22.996 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 15/142 2026-03-10T11:28:23.002 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 16/142 2026-03-10T11:28:23.009 INFO:teuthology.orchestra.run.vm06.stdout: Installing : liboath-2.6.12-1.el9.x86_64 17/142 2026-03-10T11:28:23.014 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 18/142 2026-03-10T11:28:23.042 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 19/142 2026-03-10T11:28:23.051 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 20/142 2026-03-10T11:28:23.062 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 21/142 2026-03-10T11:28:23.068 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 22/142 2026-03-10T11:28:23.072 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-5.4.4-4.el9.x86_64 23/142 2026-03-10T11:28:23.079 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 24/142 2026-03-10T11:28:23.111 INFO:teuthology.orchestra.run.vm06.stdout: Installing : unzip-6.0-59.el9.x86_64 25/142 2026-03-10T11:28:23.127 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 26/142 2026-03-10T11:28:23.132 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 27/142 2026-03-10T11:28:23.141 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 28/142 2026-03-10T11:28:23.143 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 29/142 2026-03-10T11:28:23.178 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 30/142 2026-03-10T11:28:23.185 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 31/142 2026-03-10T11:28:23.195 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 32/142 2026-03-10T11:28:23.208 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 33/142 2026-03-10T11:28:23.216 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 34/142 2026-03-10T11:28:23.245 INFO:teuthology.orchestra.run.vm06.stdout: Installing : zip-3.0-35.el9.x86_64 35/142 2026-03-10T11:28:23.250 INFO:teuthology.orchestra.run.vm06.stdout: Installing : luarocks-3.9.2-5.el9.noarch 36/142 2026-03-10T11:28:23.258 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 37/142 2026-03-10T11:28:23.289 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 38/142 2026-03-10T11:28:23.352 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 39/142 2026-03-10T11:28:23.369 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 40/142 2026-03-10T11:28:23.379 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rsa-4.9-2.el9.noarch 41/142 2026-03-10T11:28:23.384 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 42/142 2026-03-10T11:28:23.391 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 43/142 2026-03-10T11:28:23.401 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 44/142 2026-03-10T11:28:23.410 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 45/142 2026-03-10T11:28:23.415 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 46/142 2026-03-10T11:28:23.433 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 47/142 2026-03-10T11:28:23.461 INFO:teuthology.orchestra.run.vm09.stdout:(29/140): ceph-mgr-dashboard-19.2.3-678.ge911bd 5.2 MB/s | 3.8 MB 00:00 2026-03-10T11:28:23.464 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 48/142 2026-03-10T11:28:23.472 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 49/142 2026-03-10T11:28:23.478 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 50/142 2026-03-10T11:28:23.492 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 51/142 2026-03-10T11:28:23.504 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 52/142 2026-03-10T11:28:23.517 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 53/142 2026-03-10T11:28:23.584 INFO:teuthology.orchestra.run.vm09.stdout:(30/140): ceph-mgr-modules-core-19.2.3-678.ge91 2.0 MB/s | 253 kB 00:00 2026-03-10T11:28:23.584 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 54/142 2026-03-10T11:28:23.594 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 55/142 2026-03-10T11:28:23.605 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 56/142 2026-03-10T11:28:23.654 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 57/142 2026-03-10T11:28:23.705 INFO:teuthology.orchestra.run.vm09.stdout:(31/140): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 407 kB/s | 49 kB 00:00 2026-03-10T11:28:23.825 INFO:teuthology.orchestra.run.vm09.stdout:(32/140): ceph-prometheus-alerts-19.2.3-678.ge9 140 kB/s | 17 kB 00:00 2026-03-10T11:28:23.949 INFO:teuthology.orchestra.run.vm09.stdout:(33/140): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 299 kB 00:00 2026-03-10T11:28:24.044 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 58/142 2026-03-10T11:28:24.061 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 59/142 2026-03-10T11:28:24.067 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 60/142 2026-03-10T11:28:24.076 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 61/142 2026-03-10T11:28:24.084 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 62/142 2026-03-10T11:28:24.090 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 63/142 2026-03-10T11:28:24.094 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 64/142 2026-03-10T11:28:24.103 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 65/142 2026-03-10T11:28:24.112 INFO:teuthology.orchestra.run.vm09.stdout:(34/140): ceph-mgr-diskprediction-local-19.2.3- 5.8 MB/s | 7.4 MB 00:01 2026-03-10T11:28:24.114 INFO:teuthology.orchestra.run.vm06.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 66/142 2026-03-10T11:28:24.118 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 67/142 2026-03-10T11:28:24.148 INFO:teuthology.orchestra.run.vm06.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 68/142 2026-03-10T11:28:24.199 INFO:teuthology.orchestra.run.vm06.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 69/142 2026-03-10T11:28:24.214 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 70/142 2026-03-10T11:28:24.269 INFO:teuthology.orchestra.run.vm06.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 71/142 2026-03-10T11:28:24.305 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-py-1.10.0-6.el9.noarch 72/142 2026-03-10T11:28:24.310 INFO:teuthology.orchestra.run.vm09.stdout:(35/140): cephadm-19.2.3-678.ge911bdeb.el9.noar 2.1 MB/s | 769 kB 00:00 2026-03-10T11:28:24.321 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 73/142 2026-03-10T11:28:24.331 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 74/142 2026-03-10T11:28:24.338 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pluggy-0.13.1-7.el9.noarch 75/142 2026-03-10T11:28:24.380 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-iniconfig-1.1.1-7.el9.noarch 76/142 2026-03-10T11:28:24.623 INFO:teuthology.orchestra.run.vm09.stdout:(36/140): ledmon-libs-1.1.0-3.el9.x86_64.rpm 129 kB/s | 40 kB 00:00 2026-03-10T11:28:24.642 INFO:teuthology.orchestra.run.vm09.stdout:(37/140): cryptsetup-2.8.1-3.el9.x86_64.rpm 662 kB/s | 351 kB 00:00 2026-03-10T11:28:24.649 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 77/142 2026-03-10T11:28:24.682 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 78/142 2026-03-10T11:28:24.689 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 79/142 2026-03-10T11:28:24.753 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-0.3.29-1.el9.x86_64 80/142 2026-03-10T11:28:24.756 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 81/142 2026-03-10T11:28:24.778 INFO:teuthology.orchestra.run.vm09.stdout:(38/140): libconfig-1.7.2-9.el9.x86_64.rpm 463 kB/s | 72 kB 00:00 2026-03-10T11:28:24.780 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 82/142 2026-03-10T11:28:24.823 INFO:teuthology.orchestra.run.vm09.stdout:(39/140): libgfortran-11.5.0-14.el9.x86_64.rpm 4.3 MB/s | 794 kB 00:00 2026-03-10T11:28:24.908 INFO:teuthology.orchestra.run.vm09.stdout:(40/140): mailcap-2.1.49-5.el9.noarch.rpm 391 kB/s | 33 kB 00:00 2026-03-10T11:28:24.934 INFO:teuthology.orchestra.run.vm09.stdout:(41/140): libquadmath-11.5.0-14.el9.x86_64.rpm 1.2 MB/s | 184 kB 00:00 2026-03-10T11:28:24.991 INFO:teuthology.orchestra.run.vm09.stdout:(42/140): pciutils-3.7.0-7.el9.x86_64.rpm 1.1 MB/s | 93 kB 00:00 2026-03-10T11:28:25.088 INFO:teuthology.orchestra.run.vm09.stdout:(43/140): python3-cffi-1.14.5-5.el9.x86_64.rpm 1.6 MB/s | 253 kB 00:00 2026-03-10T11:28:25.155 INFO:teuthology.orchestra.run.vm09.stdout:(44/140): python3-cryptography-36.0.1-5.el9.x86 7.7 MB/s | 1.2 MB 00:00 2026-03-10T11:28:25.167 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 83/142 2026-03-10T11:28:25.175 INFO:teuthology.orchestra.run.vm09.stdout:(45/140): python3-ply-3.11-14.el9.noarch.rpm 1.2 MB/s | 106 kB 00:00 2026-03-10T11:28:25.241 INFO:teuthology.orchestra.run.vm09.stdout:(46/140): python3-pycparser-2.20-6.el9.noarch.r 1.5 MB/s | 135 kB 00:00 2026-03-10T11:28:25.259 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 84/142 2026-03-10T11:28:25.260 INFO:teuthology.orchestra.run.vm09.stdout:(47/140): python3-requests-2.25.1-10.el9.noarch 1.5 MB/s | 126 kB 00:00 2026-03-10T11:28:25.330 INFO:teuthology.orchestra.run.vm09.stdout:(48/140): python3-urllib3-1.26.5-7.el9.noarch.r 2.4 MB/s | 218 kB 00:00 2026-03-10T11:28:25.360 INFO:teuthology.orchestra.run.vm09.stdout:(49/140): unzip-6.0-59.el9.x86_64.rpm 1.8 MB/s | 182 kB 00:00 2026-03-10T11:28:25.419 INFO:teuthology.orchestra.run.vm09.stdout:(50/140): zip-3.0-35.el9.x86_64.rpm 2.9 MB/s | 266 kB 00:00 2026-03-10T11:28:25.508 INFO:teuthology.orchestra.run.vm09.stdout:(51/140): flexiblas-3.0.4-9.el9.x86_64.rpm 334 kB/s | 30 kB 00:00 2026-03-10T11:28:25.551 INFO:teuthology.orchestra.run.vm09.stdout:(52/140): boost-program-options-1.75.0-13.el9.x 544 kB/s | 104 kB 00:00 2026-03-10T11:28:25.582 INFO:teuthology.orchestra.run.vm09.stdout:(53/140): flexiblas-openblas-openmp-3.0.4-9.el9 488 kB/s | 15 kB 00:00 2026-03-10T11:28:25.642 INFO:teuthology.orchestra.run.vm09.stdout:(54/140): libnbd-1.20.3-4.el9.x86_64.rpm 2.7 MB/s | 164 kB 00:00 2026-03-10T11:28:25.674 INFO:teuthology.orchestra.run.vm09.stdout:(55/140): libpmemobj-1.12.1-1.el9.x86_64.rpm 4.9 MB/s | 160 kB 00:00 2026-03-10T11:28:25.705 INFO:teuthology.orchestra.run.vm09.stdout:(56/140): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.4 MB/s | 45 kB 00:00 2026-03-10T11:28:25.726 INFO:teuthology.orchestra.run.vm09.stdout:(57/140): flexiblas-netlib-3.0.4-9.el9.x86_64.r 14 MB/s | 3.0 MB 00:00 2026-03-10T11:28:25.761 INFO:teuthology.orchestra.run.vm09.stdout:(58/140): libstoragemgmt-1.10.1-1.el9.x86_64.rp 7.1 MB/s | 246 kB 00:00 2026-03-10T11:28:25.770 INFO:teuthology.orchestra.run.vm09.stdout:(59/140): librdkafka-1.6.1-102.el9.x86_64.rpm 10 MB/s | 662 kB 00:00 2026-03-10T11:28:25.927 INFO:teuthology.orchestra.run.vm09.stdout:(60/140): ceph-test-19.2.3-678.ge911bdeb.el9.x8 8.6 MB/s | 50 MB 00:05 2026-03-10T11:28:25.929 INFO:teuthology.orchestra.run.vm09.stdout:(61/140): libxslt-1.1.34-12.el9.x86_64.rpm 1.4 MB/s | 233 kB 00:00 2026-03-10T11:28:25.930 INFO:teuthology.orchestra.run.vm09.stdout:(62/140): lttng-ust-2.12.0-6.el9.x86_64.rpm 1.8 MB/s | 292 kB 00:00 2026-03-10T11:28:25.960 INFO:teuthology.orchestra.run.vm09.stdout:(63/140): openblas-0.3.29-1.el9.x86_64.rpm 1.4 MB/s | 42 kB 00:00 2026-03-10T11:28:26.058 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 85/142 2026-03-10T11:28:26.075 INFO:teuthology.orchestra.run.vm09.stdout:(64/140): lua-5.4.4-4.el9.x86_64.rpm 1.2 MB/s | 188 kB 00:00 2026-03-10T11:28:26.087 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 86/142 2026-03-10T11:28:26.095 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 87/142 2026-03-10T11:28:26.100 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 88/142 2026-03-10T11:28:26.131 INFO:teuthology.orchestra.run.vm09.stdout:(65/140): openblas-openmp-0.3.29-1.el9.x86_64.r 26 MB/s | 5.3 MB 00:00 2026-03-10T11:28:26.194 INFO:teuthology.orchestra.run.vm09.stdout:(66/140): python3-devel-3.9.25-3.el9.x86_64.rpm 3.8 MB/s | 244 kB 00:00 2026-03-10T11:28:26.251 INFO:teuthology.orchestra.run.vm09.stdout:(67/140): python3-iniconfig-1.1.1-7.el9.noarch. 306 kB/s | 17 kB 00:00 2026-03-10T11:28:26.262 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 89/142 2026-03-10T11:28:26.265 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 90/142 2026-03-10T11:28:26.285 INFO:teuthology.orchestra.run.vm09.stdout:(68/140): python3-jinja2-2.11.3-8.el9.noarch.rp 7.2 MB/s | 249 kB 00:00 2026-03-10T11:28:26.300 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 90/142 2026-03-10T11:28:26.304 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 91/142 2026-03-10T11:28:26.312 INFO:teuthology.orchestra.run.vm06.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 92/142 2026-03-10T11:28:26.317 INFO:teuthology.orchestra.run.vm09.stdout:(69/140): python3-jmespath-1.0.1-1.el9.noarch.r 1.5 MB/s | 48 kB 00:00 2026-03-10T11:28:26.364 INFO:teuthology.orchestra.run.vm09.stdout:(70/140): python3-libstoragemgmt-1.10.1-1.el9.x 3.7 MB/s | 177 kB 00:00 2026-03-10T11:28:26.367 INFO:teuthology.orchestra.run.vm09.stdout:(71/140): protobuf-3.14.0-17.el9.x86_64.rpm 2.5 MB/s | 1.0 MB 00:00 2026-03-10T11:28:26.435 INFO:teuthology.orchestra.run.vm09.stdout:(72/140): python3-mako-1.1.4-6.el9.noarch.rpm 2.4 MB/s | 172 kB 00:00 2026-03-10T11:28:26.486 INFO:teuthology.orchestra.run.vm09.stdout:(73/140): python3-markupsafe-1.1.1-12.el9.x86_6 292 kB/s | 35 kB 00:00 2026-03-10T11:28:26.546 INFO:teuthology.orchestra.run.vm09.stdout:(74/140): python3-numpy-f2py-1.23.5-2.el9.x86_6 7.3 MB/s | 442 kB 00:00 2026-03-10T11:28:26.573 INFO:teuthology.orchestra.run.vm06.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 93/142 2026-03-10T11:28:26.590 INFO:teuthology.orchestra.run.vm09.stdout:(75/140): python3-packaging-20.9-5.el9.noarch.r 1.7 MB/s | 77 kB 00:00 2026-03-10T11:28:26.590 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 94/142 2026-03-10T11:28:26.615 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 94/142 2026-03-10T11:28:26.617 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 95/142 2026-03-10T11:28:26.645 INFO:teuthology.orchestra.run.vm09.stdout:(76/140): python3-babel-2.9.1-2.el9.noarch.rpm 10 MB/s | 6.0 MB 00:00 2026-03-10T11:28:26.646 INFO:teuthology.orchestra.run.vm09.stdout:(77/140): python3-pluggy-0.13.1-7.el9.noarch.rp 738 kB/s | 41 kB 00:00 2026-03-10T11:28:26.675 INFO:teuthology.orchestra.run.vm09.stdout:(78/140): python3-numpy-1.23.5-2.el9.x86_64.rpm 26 MB/s | 6.1 MB 00:00 2026-03-10T11:28:26.679 INFO:teuthology.orchestra.run.vm09.stdout:(79/140): python3-protobuf-3.14.0-17.el9.noarch 7.7 MB/s | 267 kB 00:00 2026-03-10T11:28:26.702 INFO:teuthology.orchestra.run.vm09.stdout:(80/140): python3-py-1.10.0-6.el9.noarch.rpm 8.4 MB/s | 477 kB 00:00 2026-03-10T11:28:26.707 INFO:teuthology.orchestra.run.vm09.stdout:(81/140): python3-pyasn1-0.4.8-7.el9.noarch.rpm 4.8 MB/s | 157 kB 00:00 2026-03-10T11:28:26.712 INFO:teuthology.orchestra.run.vm09.stdout:(82/140): python3-pyasn1-modules-0.4.8-7.el9.no 8.1 MB/s | 277 kB 00:00 2026-03-10T11:28:26.761 INFO:teuthology.orchestra.run.vm09.stdout:(83/140): python3-pytest-6.2.2-7.el9.noarch.rpm 8.7 MB/s | 519 kB 00:00 2026-03-10T11:28:26.761 INFO:teuthology.orchestra.run.vm09.stdout:(84/140): python3-requests-oauthlib-1.3.0-12.el 993 kB/s | 54 kB 00:00 2026-03-10T11:28:26.793 INFO:teuthology.orchestra.run.vm09.stdout:(85/140): python3-toml-0.10.2-6.el9.noarch.rpm 1.2 MB/s | 42 kB 00:00 2026-03-10T11:28:26.796 INFO:teuthology.orchestra.run.vm09.stdout:(86/140): qatlib-25.08.0-2.el9.x86_64.rpm 6.7 MB/s | 240 kB 00:00 2026-03-10T11:28:26.824 INFO:teuthology.orchestra.run.vm09.stdout:(87/140): qatlib-service-25.08.0-2.el9.x86_64.r 1.2 MB/s | 37 kB 00:00 2026-03-10T11:28:26.828 INFO:teuthology.orchestra.run.vm09.stdout:(88/140): qatzip-libs-1.3.1-1.el9.x86_64.rpm 2.1 MB/s | 66 kB 00:00 2026-03-10T11:28:26.860 INFO:teuthology.orchestra.run.vm09.stdout:(89/140): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.9 MB/s | 64 kB 00:00 2026-03-10T11:28:26.885 INFO:teuthology.orchestra.run.vm09.stdout:(90/140): socat-1.7.4.1-8.el9.x86_64.rpm 4.8 MB/s | 303 kB 00:00 2026-03-10T11:28:26.944 INFO:teuthology.orchestra.run.vm09.stdout:(91/140): lua-devel-5.4.4-4.el9.x86_64.rpm 266 kB/s | 22 kB 00:00 2026-03-10T11:28:26.959 INFO:teuthology.orchestra.run.vm09.stdout:(92/140): abseil-cpp-20211102.0-4.el9.x86_64.rp 37 MB/s | 551 kB 00:00 2026-03-10T11:28:26.965 INFO:teuthology.orchestra.run.vm09.stdout:(93/140): gperftools-libs-2.9.1-3.el9.x86_64.rp 54 MB/s | 308 kB 00:00 2026-03-10T11:28:26.967 INFO:teuthology.orchestra.run.vm09.stdout:(94/140): grpc-data-1.46.7-10.el9.noarch.rpm 7.4 MB/s | 19 kB 00:00 2026-03-10T11:28:27.038 INFO:teuthology.orchestra.run.vm09.stdout:(95/140): libarrow-9.0.0-15.el9.x86_64.rpm 62 MB/s | 4.4 MB 00:00 2026-03-10T11:28:27.041 INFO:teuthology.orchestra.run.vm09.stdout:(96/140): libarrow-doc-9.0.0-15.el9.noarch.rpm 10 MB/s | 25 kB 00:00 2026-03-10T11:28:27.043 INFO:teuthology.orchestra.run.vm09.stdout:(97/140): liboath-2.6.12-1.el9.x86_64.rpm 20 MB/s | 49 kB 00:00 2026-03-10T11:28:27.047 INFO:teuthology.orchestra.run.vm09.stdout:(98/140): libunwind-1.6.2-1.el9.x86_64.rpm 22 MB/s | 67 kB 00:00 2026-03-10T11:28:27.051 INFO:teuthology.orchestra.run.vm09.stdout:(99/140): luarocks-3.9.2-5.el9.noarch.rpm 37 MB/s | 151 kB 00:00 2026-03-10T11:28:27.063 INFO:teuthology.orchestra.run.vm09.stdout:(100/140): parquet-libs-9.0.0-15.el9.x86_64.rpm 72 MB/s | 838 kB 00:00 2026-03-10T11:28:27.072 INFO:teuthology.orchestra.run.vm09.stdout:(101/140): python3-asyncssh-2.13.2-5.el9.noarch 60 MB/s | 548 kB 00:00 2026-03-10T11:28:27.078 INFO:teuthology.orchestra.run.vm09.stdout:(102/140): python3-autocommand-2.2.2-8.el9.noar 5.1 MB/s | 29 kB 00:00 2026-03-10T11:28:27.081 INFO:teuthology.orchestra.run.vm09.stdout:(103/140): python3-backports-tarfile-1.2.0-1.el 17 MB/s | 60 kB 00:00 2026-03-10T11:28:27.084 INFO:teuthology.orchestra.run.vm09.stdout:(104/140): python3-bcrypt-3.2.2-1.el9.x86_64.rp 17 MB/s | 43 kB 00:00 2026-03-10T11:28:27.086 INFO:teuthology.orchestra.run.vm09.stdout:(105/140): python3-cachetools-4.2.4-1.el9.noarc 14 MB/s | 32 kB 00:00 2026-03-10T11:28:27.088 INFO:teuthology.orchestra.run.vm09.stdout:(106/140): python3-certifi-2023.05.07-4.el9.noa 7.3 MB/s | 14 kB 00:00 2026-03-10T11:28:27.093 INFO:teuthology.orchestra.run.vm09.stdout:(107/140): python3-cheroot-10.0.1-4.el9.noarch. 39 MB/s | 173 kB 00:00 2026-03-10T11:28:27.099 INFO:teuthology.orchestra.run.vm09.stdout:(108/140): python3-cherrypy-18.6.1-2.el9.noarch 60 MB/s | 358 kB 00:00 2026-03-10T11:28:27.104 INFO:teuthology.orchestra.run.vm09.stdout:(109/140): python3-google-auth-2.45.0-1.el9.noa 53 MB/s | 254 kB 00:00 2026-03-10T11:28:27.120 INFO:teuthology.orchestra.run.vm09.stdout:(110/140): protobuf-compiler-3.14.0-17.el9.x86_ 3.6 MB/s | 862 kB 00:00 2026-03-10T11:28:27.137 INFO:teuthology.orchestra.run.vm09.stdout:(111/140): python3-grpcio-1.46.7-10.el9.x86_64. 62 MB/s | 2.0 MB 00:00 2026-03-10T11:28:27.138 INFO:teuthology.orchestra.run.vm09.stdout:(112/140): python3-grpcio-tools-1.46.7-10.el9.x 7.6 MB/s | 144 kB 00:00 2026-03-10T11:28:27.139 INFO:teuthology.orchestra.run.vm09.stdout:(113/140): python3-jaraco-8.2.1-3.el9.noarch.rp 4.8 MB/s | 11 kB 00:00 2026-03-10T11:28:27.141 INFO:teuthology.orchestra.run.vm09.stdout:(114/140): python3-jaraco-classes-3.2.1-5.el9.n 8.9 MB/s | 18 kB 00:00 2026-03-10T11:28:27.142 INFO:teuthology.orchestra.run.vm09.stdout:(115/140): python3-jaraco-collections-3.0.0-8.e 8.5 MB/s | 23 kB 00:00 2026-03-10T11:28:27.143 INFO:teuthology.orchestra.run.vm09.stdout:(116/140): python3-jaraco-context-6.0.1-3.el9.n 9.7 MB/s | 20 kB 00:00 2026-03-10T11:28:27.144 INFO:teuthology.orchestra.run.vm09.stdout:(117/140): python3-jaraco-functools-3.5.0-2.el9 10 MB/s | 19 kB 00:00 2026-03-10T11:28:27.145 INFO:teuthology.orchestra.run.vm09.stdout:(118/140): python3-jaraco-text-4.0.0-2.el9.noar 11 MB/s | 26 kB 00:00 2026-03-10T11:28:27.151 INFO:teuthology.orchestra.run.vm09.stdout:(119/140): python3-logutils-0.3.5-21.el9.noarch 8.1 MB/s | 46 kB 00:00 2026-03-10T11:28:27.155 INFO:teuthology.orchestra.run.vm09.stdout:(120/140): python3-more-itertools-8.12.0-2.el9. 19 MB/s | 79 kB 00:00 2026-03-10T11:28:27.159 INFO:teuthology.orchestra.run.vm09.stdout:(121/140): python3-kubernetes-26.1.0-3.el9.noar 67 MB/s | 1.0 MB 00:00 2026-03-10T11:28:27.161 INFO:teuthology.orchestra.run.vm09.stdout:(122/140): python3-natsort-7.1.1-5.el9.noarch.r 11 MB/s | 58 kB 00:00 2026-03-10T11:28:27.165 INFO:teuthology.orchestra.run.vm09.stdout:(123/140): python3-pecan-1.4.2-3.el9.noarch.rpm 53 MB/s | 272 kB 00:00 2026-03-10T11:28:27.165 INFO:teuthology.orchestra.run.vm09.stdout:(124/140): python3-portend-3.1.0-2.el9.noarch.r 3.7 MB/s | 16 kB 00:00 2026-03-10T11:28:27.167 INFO:teuthology.orchestra.run.vm09.stdout:(125/140): python3-pyOpenSSL-21.0.0-1.el9.noarc 32 MB/s | 90 kB 00:00 2026-03-10T11:28:27.169 INFO:teuthology.orchestra.run.vm09.stdout:(126/140): python3-repoze-lru-0.7-16.el9.noarch 7.5 MB/s | 31 kB 00:00 2026-03-10T11:28:27.172 INFO:teuthology.orchestra.run.vm09.stdout:(127/140): python3-routes-2.5.1-5.el9.noarch.rp 38 MB/s | 188 kB 00:00 2026-03-10T11:28:27.173 INFO:teuthology.orchestra.run.vm09.stdout:(128/140): python3-rsa-4.9-2.el9.noarch.rpm 15 MB/s | 59 kB 00:00 2026-03-10T11:28:27.175 INFO:teuthology.orchestra.run.vm09.stdout:(129/140): python3-tempora-5.0.0-2.el9.noarch.r 17 MB/s | 36 kB 00:00 2026-03-10T11:28:27.176 INFO:teuthology.orchestra.run.vm09.stdout:(130/140): python3-typing-extensions-4.15.0-1.e 32 MB/s | 86 kB 00:00 2026-03-10T11:28:27.182 INFO:teuthology.orchestra.run.vm09.stdout:(131/140): python3-webob-1.8.8-2.el9.noarch.rpm 38 MB/s | 230 kB 00:00 2026-03-10T11:28:27.183 INFO:teuthology.orchestra.run.vm09.stdout:(132/140): python3-websocket-client-1.2.3-2.el9 14 MB/s | 90 kB 00:00 2026-03-10T11:28:27.187 INFO:teuthology.orchestra.run.vm09.stdout:(133/140): python3-xmltodict-0.12.0-15.el9.noar 5.9 MB/s | 22 kB 00:00 2026-03-10T11:28:27.188 INFO:teuthology.orchestra.run.vm09.stdout:(134/140): python3-werkzeug-2.0.3-3.el9.1.noarc 60 MB/s | 427 kB 00:00 2026-03-10T11:28:27.189 INFO:teuthology.orchestra.run.vm09.stdout:(135/140): python3-zc-lockfile-2.0-10.el9.noarc 9.7 MB/s | 20 kB 00:00 2026-03-10T11:28:27.192 INFO:teuthology.orchestra.run.vm09.stdout:(136/140): re2-20211101-20.el9.x86_64.rpm 50 MB/s | 191 kB 00:00 2026-03-10T11:28:27.217 INFO:teuthology.orchestra.run.vm09.stdout:(137/140): thrift-0.15.0-4.el9.x86_64.rpm 56 MB/s | 1.6 MB 00:00 2026-03-10T11:28:27.512 INFO:teuthology.orchestra.run.vm09.stdout:(138/140): python3-scipy-1.9.3-2.el9.x86_64.rpm 24 MB/s | 19 MB 00:00 2026-03-10T11:28:27.742 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 96/142 2026-03-10T11:28:27.748 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 96/142 2026-03-10T11:28:27.770 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 96/142 2026-03-10T11:28:27.787 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ply-3.11-14.el9.noarch 97/142 2026-03-10T11:28:27.808 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 98/142 2026-03-10T11:28:27.899 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 99/142 2026-03-10T11:28:27.932 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 100/142 2026-03-10T11:28:27.963 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 101/142 2026-03-10T11:28:27.999 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 102/142 2026-03-10T11:28:28.040 INFO:teuthology.orchestra.run.vm09.stdout:(139/140): librados2-19.2.3-678.ge911bdeb.el9.x 4.0 MB/s | 3.4 MB 00:00 2026-03-10T11:28:28.062 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 103/142 2026-03-10T11:28:28.072 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 104/142 2026-03-10T11:28:28.078 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 105/142 2026-03-10T11:28:28.084 INFO:teuthology.orchestra.run.vm06.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 106/142 2026-03-10T11:28:28.088 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 107/142 2026-03-10T11:28:28.090 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 108/142 2026-03-10T11:28:28.107 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 108/142 2026-03-10T11:28:28.361 INFO:teuthology.orchestra.run.vm09.stdout:(140/140): librbd1-19.2.3-678.ge911bdeb.el9.x86 2.8 MB/s | 3.2 MB 00:01 2026-03-10T11:28:28.364 INFO:teuthology.orchestra.run.vm09.stdout:-------------------------------------------------------------------------------- 2026-03-10T11:28:28.364 INFO:teuthology.orchestra.run.vm09.stdout:Total 16 MB/s | 211 MB 00:13 2026-03-10T11:28:28.403 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 109/142 2026-03-10T11:28:28.409 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 110/142 2026-03-10T11:28:28.458 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 110/142 2026-03-10T11:28:28.458 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-10T11:28:28.458 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-10T11:28:28.458 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:28.464 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 111/142 2026-03-10T11:28:28.923 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T11:28:28.972 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T11:28:28.972 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T11:28:29.793 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T11:28:29.793 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T11:28:30.723 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T11:28:30.737 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/142 2026-03-10T11:28:30.750 INFO:teuthology.orchestra.run.vm09.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/142 2026-03-10T11:28:30.922 INFO:teuthology.orchestra.run.vm09.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/142 2026-03-10T11:28:30.924 INFO:teuthology.orchestra.run.vm09.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/142 2026-03-10T11:28:30.987 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/142 2026-03-10T11:28:30.989 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/142 2026-03-10T11:28:31.018 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/142 2026-03-10T11:28:31.028 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/142 2026-03-10T11:28:31.032 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/142 2026-03-10T11:28:31.034 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/142 2026-03-10T11:28:31.045 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/142 2026-03-10T11:28:31.052 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-packaging-20.9-5.el9.noarch 10/142 2026-03-10T11:28:31.062 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 11/142 2026-03-10T11:28:31.063 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 12/142 2026-03-10T11:28:31.099 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 12/142 2026-03-10T11:28:31.101 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 13/142 2026-03-10T11:28:31.118 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 13/142 2026-03-10T11:28:31.157 INFO:teuthology.orchestra.run.vm09.stdout: Installing : re2-1:20211101-20.el9.x86_64 14/142 2026-03-10T11:28:31.199 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 15/142 2026-03-10T11:28:31.205 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 16/142 2026-03-10T11:28:31.212 INFO:teuthology.orchestra.run.vm09.stdout: Installing : liboath-2.6.12-1.el9.x86_64 17/142 2026-03-10T11:28:31.217 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 18/142 2026-03-10T11:28:31.245 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 19/142 2026-03-10T11:28:31.254 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 20/142 2026-03-10T11:28:31.265 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 21/142 2026-03-10T11:28:31.273 INFO:teuthology.orchestra.run.vm09.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 22/142 2026-03-10T11:28:31.278 INFO:teuthology.orchestra.run.vm09.stdout: Installing : lua-5.4.4-4.el9.x86_64 23/142 2026-03-10T11:28:31.283 INFO:teuthology.orchestra.run.vm09.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 24/142 2026-03-10T11:28:31.314 INFO:teuthology.orchestra.run.vm09.stdout: Installing : unzip-6.0-59.el9.x86_64 25/142 2026-03-10T11:28:31.331 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 26/142 2026-03-10T11:28:31.336 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 27/142 2026-03-10T11:28:31.344 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 28/142 2026-03-10T11:28:31.347 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 29/142 2026-03-10T11:28:31.380 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 30/142 2026-03-10T11:28:31.388 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 31/142 2026-03-10T11:28:31.399 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 32/142 2026-03-10T11:28:31.414 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 33/142 2026-03-10T11:28:31.423 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 34/142 2026-03-10T11:28:31.457 INFO:teuthology.orchestra.run.vm09.stdout: Installing : zip-3.0-35.el9.x86_64 35/142 2026-03-10T11:28:31.463 INFO:teuthology.orchestra.run.vm09.stdout: Installing : luarocks-3.9.2-5.el9.noarch 36/142 2026-03-10T11:28:31.473 INFO:teuthology.orchestra.run.vm09.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 37/142 2026-03-10T11:28:31.504 INFO:teuthology.orchestra.run.vm09.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 38/142 2026-03-10T11:28:31.568 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 39/142 2026-03-10T11:28:31.585 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 40/142 2026-03-10T11:28:31.594 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rsa-4.9-2.el9.noarch 41/142 2026-03-10T11:28:31.600 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 42/142 2026-03-10T11:28:31.607 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 43/142 2026-03-10T11:28:31.616 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 44/142 2026-03-10T11:28:31.623 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 45/142 2026-03-10T11:28:31.627 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 46/142 2026-03-10T11:28:31.646 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 47/142 2026-03-10T11:28:31.674 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 48/142 2026-03-10T11:28:31.682 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 49/142 2026-03-10T11:28:31.689 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 50/142 2026-03-10T11:28:31.707 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 51/142 2026-03-10T11:28:31.720 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 52/142 2026-03-10T11:28:31.732 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 53/142 2026-03-10T11:28:31.802 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 54/142 2026-03-10T11:28:31.979 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 55/142 2026-03-10T11:28:32.088 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 56/142 2026-03-10T11:28:32.237 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 57/142 2026-03-10T11:28:32.654 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 58/142 2026-03-10T11:28:32.672 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 59/142 2026-03-10T11:28:32.680 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 60/142 2026-03-10T11:28:32.689 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 61/142 2026-03-10T11:28:32.699 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 62/142 2026-03-10T11:28:32.705 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 63/142 2026-03-10T11:28:32.712 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 64/142 2026-03-10T11:28:32.721 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 65/142 2026-03-10T11:28:32.725 INFO:teuthology.orchestra.run.vm09.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 66/142 2026-03-10T11:28:32.728 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 67/142 2026-03-10T11:28:32.758 INFO:teuthology.orchestra.run.vm09.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 68/142 2026-03-10T11:28:32.809 INFO:teuthology.orchestra.run.vm09.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 69/142 2026-03-10T11:28:32.822 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 70/142 2026-03-10T11:28:32.876 INFO:teuthology.orchestra.run.vm09.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 71/142 2026-03-10T11:28:32.912 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-py-1.10.0-6.el9.noarch 72/142 2026-03-10T11:28:32.925 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 73/142 2026-03-10T11:28:32.935 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 74/142 2026-03-10T11:28:32.941 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pluggy-0.13.1-7.el9.noarch 75/142 2026-03-10T11:28:32.983 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-iniconfig-1.1.1-7.el9.noarch 76/142 2026-03-10T11:28:33.263 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 77/142 2026-03-10T11:28:33.296 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 78/142 2026-03-10T11:28:33.303 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 79/142 2026-03-10T11:28:33.364 INFO:teuthology.orchestra.run.vm09.stdout: Installing : openblas-0.3.29-1.el9.x86_64 80/142 2026-03-10T11:28:33.368 INFO:teuthology.orchestra.run.vm09.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 81/142 2026-03-10T11:28:33.395 INFO:teuthology.orchestra.run.vm09.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 82/142 2026-03-10T11:28:33.796 INFO:teuthology.orchestra.run.vm09.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 83/142 2026-03-10T11:28:34.103 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 84/142 2026-03-10T11:28:35.102 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 111/142 2026-03-10T11:28:35.102 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-03-10T11:28:35.102 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-03-10T11:28:35.102 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-03-10T11:28:35.102 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-03-10T11:28:35.102 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-03-10T11:28:35.102 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-03-10T11:28:35.102 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-03-10T11:28:35.102 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:35.164 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 85/142 2026-03-10T11:28:35.193 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 86/142 2026-03-10T11:28:35.199 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 87/142 2026-03-10T11:28:35.205 INFO:teuthology.orchestra.run.vm09.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 88/142 2026-03-10T11:28:35.222 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 112/142 2026-03-10T11:28:35.247 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 112/142 2026-03-10T11:28:35.247 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:35.247 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T11:28:35.247 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T11:28:35.247 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T11:28:35.247 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:35.366 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 89/142 2026-03-10T11:28:35.369 INFO:teuthology.orchestra.run.vm09.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 90/142 2026-03-10T11:28:35.403 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 90/142 2026-03-10T11:28:35.411 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 91/142 2026-03-10T11:28:35.419 INFO:teuthology.orchestra.run.vm09.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 92/142 2026-03-10T11:28:35.479 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 113/142 2026-03-10T11:28:35.504 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 113/142 2026-03-10T11:28:35.504 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:35.504 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T11:28:35.504 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T11:28:35.504 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T11:28:35.504 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:35.512 INFO:teuthology.orchestra.run.vm06.stdout: Installing : mailcap-2.1.49-5.el9.noarch 114/142 2026-03-10T11:28:35.515 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 115/142 2026-03-10T11:28:35.535 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 116/142 2026-03-10T11:28:35.535 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'qat' with GID 994. 2026-03-10T11:28:35.535 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-10T11:28:35.535 INFO:teuthology.orchestra.run.vm06.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-10T11:28:35.535 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:35.546 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 116/142 2026-03-10T11:28:35.573 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 116/142 2026-03-10T11:28:35.573 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-10T11:28:35.573 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:35.615 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 117/142 2026-03-10T11:28:35.675 INFO:teuthology.orchestra.run.vm09.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 93/142 2026-03-10T11:28:35.679 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 94/142 2026-03-10T11:28:35.688 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 118/142 2026-03-10T11:28:35.693 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 119/142 2026-03-10T11:28:35.697 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 94/142 2026-03-10T11:28:35.700 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 95/142 2026-03-10T11:28:35.707 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 119/142 2026-03-10T11:28:35.707 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:35.707 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T11:28:35.707 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:36.546 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 120/142 2026-03-10T11:28:36.572 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 120/142 2026-03-10T11:28:36.572 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:36.572 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T11:28:36.572 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T11:28:36.572 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T11:28:36.572 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:36.665 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 121/142 2026-03-10T11:28:36.672 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 121/142 2026-03-10T11:28:36.678 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 122/142 2026-03-10T11:28:36.701 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 123/142 2026-03-10T11:28:36.704 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 124/142 2026-03-10T11:28:36.881 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 96/142 2026-03-10T11:28:36.888 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 96/142 2026-03-10T11:28:36.913 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 96/142 2026-03-10T11:28:36.931 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-ply-3.11-14.el9.noarch 97/142 2026-03-10T11:28:36.952 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 98/142 2026-03-10T11:28:37.052 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 99/142 2026-03-10T11:28:37.069 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 100/142 2026-03-10T11:28:37.097 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 101/142 2026-03-10T11:28:37.138 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 102/142 2026-03-10T11:28:37.205 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 103/142 2026-03-10T11:28:37.218 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 104/142 2026-03-10T11:28:37.224 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 105/142 2026-03-10T11:28:37.231 INFO:teuthology.orchestra.run.vm09.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 106/142 2026-03-10T11:28:37.236 INFO:teuthology.orchestra.run.vm09.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 107/142 2026-03-10T11:28:37.238 INFO:teuthology.orchestra.run.vm09.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 108/142 2026-03-10T11:28:37.256 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 124/142 2026-03-10T11:28:37.257 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 108/142 2026-03-10T11:28:37.263 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 125/142 2026-03-10T11:28:37.616 INFO:teuthology.orchestra.run.vm09.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 109/142 2026-03-10T11:28:37.624 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 110/142 2026-03-10T11:28:37.672 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 110/142 2026-03-10T11:28:37.673 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-10T11:28:37.673 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-10T11:28:37.673 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:37.678 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 111/142 2026-03-10T11:28:37.816 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 125/142 2026-03-10T11:28:37.818 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 126/142 2026-03-10T11:28:37.882 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 126/142 2026-03-10T11:28:37.937 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 127/142 2026-03-10T11:28:37.939 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 128/142 2026-03-10T11:28:37.961 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 128/142 2026-03-10T11:28:37.961 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:37.961 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T11:28:37.961 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T11:28:37.961 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T11:28:37.961 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:37.975 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 129/142 2026-03-10T11:28:37.986 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 129/142 2026-03-10T11:28:38.532 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 130/142 2026-03-10T11:28:38.545 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 131/142 2026-03-10T11:28:38.567 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 131/142 2026-03-10T11:28:38.567 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:38.567 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T11:28:38.567 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T11:28:38.567 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T11:28:38.568 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:38.579 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 132/142 2026-03-10T11:28:38.601 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 132/142 2026-03-10T11:28:38.601 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:38.601 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T11:28:38.601 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:38.771 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 133/142 2026-03-10T11:28:38.792 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 133/142 2026-03-10T11:28:38.792 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:38.792 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T11:28:38.792 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T11:28:38.792 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T11:28:38.792 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:41.531 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 134/142 2026-03-10T11:28:41.542 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/142 2026-03-10T11:28:41.597 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 136/142 2026-03-10T11:28:41.605 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pytest-6.2.2-7.el9.noarch 137/142 2026-03-10T11:28:41.668 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 138/142 2026-03-10T11:28:41.678 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 139/142 2026-03-10T11:28:41.683 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 140/142 2026-03-10T11:28:41.683 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 141/142 2026-03-10T11:28:41.699 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 141/142 2026-03-10T11:28:41.699 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 142/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 142/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/142 2026-03-10T11:28:43.936 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 51/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-iniconfig-1.1.1-7.el9.noarch 69/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 70/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 71/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 72/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 73/142 2026-03-10T11:28:43.937 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 74/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 75/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 76/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 77/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pluggy-0.13.1-7.el9.noarch 78/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 79/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-py-1.10.0-6.el9.noarch 80/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 81/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 82/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pytest-6.2.2-7.el9.noarch 83/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 84/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 85/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 86/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 87/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 88/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 89/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 90/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 91/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 92/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 93/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 94/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 95/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 96/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 97/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 98/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 99/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 100/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 101/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 102/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 103/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 104/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 105/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 106/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 107/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 108/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 109/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 110/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 111/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 112/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 113/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 114/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 115/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 116/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 117/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 118/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 119/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 120/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 121/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 122/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 123/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 124/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 125/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 126/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 127/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 128/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 129/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 130/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 131/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 132/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 133/142 2026-03-10T11:28:43.938 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 134/142 2026-03-10T11:28:43.939 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 135/142 2026-03-10T11:28:43.939 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 136/142 2026-03-10T11:28:43.939 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 137/142 2026-03-10T11:28:43.939 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 138/142 2026-03-10T11:28:43.939 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 139/142 2026-03-10T11:28:43.939 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 140/142 2026-03-10T11:28:43.939 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 141/142 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 142/142 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout:Upgraded: 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T11:28:44.384 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-iniconfig-1.1.1-7.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-pluggy-0.13.1-7.el9.noarch 2026-03-10T11:28:44.385 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-py-1.10.0-6.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytest-6.2.2-7.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:28:44.386 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:28:44.469 DEBUG:teuthology.parallel:result is None 2026-03-10T11:28:44.734 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 111/142 2026-03-10T11:28:44.734 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /sys 2026-03-10T11:28:44.734 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /proc 2026-03-10T11:28:44.734 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /mnt 2026-03-10T11:28:44.734 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /var/tmp 2026-03-10T11:28:44.734 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /home 2026-03-10T11:28:44.734 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /root 2026-03-10T11:28:44.734 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /tmp 2026-03-10T11:28:44.734 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:44.855 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 112/142 2026-03-10T11:28:44.877 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 112/142 2026-03-10T11:28:44.877 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:44.877 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T11:28:44.877 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T11:28:44.877 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T11:28:44.877 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:45.099 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 113/142 2026-03-10T11:28:45.122 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 113/142 2026-03-10T11:28:45.122 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:45.122 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T11:28:45.122 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T11:28:45.122 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T11:28:45.122 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:45.131 INFO:teuthology.orchestra.run.vm09.stdout: Installing : mailcap-2.1.49-5.el9.noarch 114/142 2026-03-10T11:28:45.135 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 115/142 2026-03-10T11:28:45.154 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 116/142 2026-03-10T11:28:45.154 INFO:teuthology.orchestra.run.vm09.stdout:Creating group 'qat' with GID 994. 2026-03-10T11:28:45.154 INFO:teuthology.orchestra.run.vm09.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-10T11:28:45.154 INFO:teuthology.orchestra.run.vm09.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-10T11:28:45.154 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:45.165 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 116/142 2026-03-10T11:28:45.190 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 116/142 2026-03-10T11:28:45.190 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-10T11:28:45.190 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:45.233 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 117/142 2026-03-10T11:28:45.307 INFO:teuthology.orchestra.run.vm09.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 118/142 2026-03-10T11:28:45.313 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 119/142 2026-03-10T11:28:45.326 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 119/142 2026-03-10T11:28:45.326 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:45.326 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T11:28:45.326 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:46.109 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 120/142 2026-03-10T11:28:46.134 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 120/142 2026-03-10T11:28:46.134 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:46.134 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T11:28:46.134 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T11:28:46.134 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T11:28:46.134 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:46.191 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 121/142 2026-03-10T11:28:46.195 INFO:teuthology.orchestra.run.vm09.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 121/142 2026-03-10T11:28:46.201 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 122/142 2026-03-10T11:28:46.224 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 123/142 2026-03-10T11:28:46.227 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 124/142 2026-03-10T11:28:46.756 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 124/142 2026-03-10T11:28:46.762 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 125/142 2026-03-10T11:28:47.274 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 125/142 2026-03-10T11:28:47.277 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 126/142 2026-03-10T11:28:47.340 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 126/142 2026-03-10T11:28:47.400 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 127/142 2026-03-10T11:28:47.403 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 128/142 2026-03-10T11:28:47.426 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 128/142 2026-03-10T11:28:47.426 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:47.426 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T11:28:47.426 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T11:28:47.426 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T11:28:47.426 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:47.440 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 129/142 2026-03-10T11:28:47.451 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 129/142 2026-03-10T11:28:47.973 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 130/142 2026-03-10T11:28:47.977 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 131/142 2026-03-10T11:28:47.998 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 131/142 2026-03-10T11:28:47.998 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:47.998 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T11:28:47.998 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T11:28:47.998 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T11:28:47.998 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:48.009 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 132/142 2026-03-10T11:28:48.027 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 132/142 2026-03-10T11:28:48.027 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:48.027 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T11:28:48.027 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:48.173 INFO:teuthology.orchestra.run.vm09.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 133/142 2026-03-10T11:28:48.192 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 133/142 2026-03-10T11:28:48.192 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:28:48.192 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T11:28:48.192 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T11:28:48.192 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T11:28:48.192 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:50.744 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 134/142 2026-03-10T11:28:50.755 INFO:teuthology.orchestra.run.vm09.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/142 2026-03-10T11:28:50.807 INFO:teuthology.orchestra.run.vm09.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 136/142 2026-03-10T11:28:50.815 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pytest-6.2.2-7.el9.noarch 137/142 2026-03-10T11:28:50.872 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 138/142 2026-03-10T11:28:50.881 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 139/142 2026-03-10T11:28:50.886 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 140/142 2026-03-10T11:28:50.887 INFO:teuthology.orchestra.run.vm09.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 141/142 2026-03-10T11:28:50.903 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 141/142 2026-03-10T11:28:50.903 INFO:teuthology.orchestra.run.vm09.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 142/142 2026-03-10T11:28:52.213 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 142/142 2026-03-10T11:28:52.213 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/142 2026-03-10T11:28:52.213 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/142 2026-03-10T11:28:52.213 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/142 2026-03-10T11:28:52.213 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/142 2026-03-10T11:28:52.213 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/142 2026-03-10T11:28:52.213 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/142 2026-03-10T11:28:52.213 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/142 2026-03-10T11:28:52.214 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/142 2026-03-10T11:28:52.215 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/142 2026-03-10T11:28:52.215 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : zip-3.0-35.el9.x86_64 51/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/142 2026-03-10T11:28:52.216 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-iniconfig-1.1.1-7.el9.noarch 69/142 2026-03-10T11:28:52.217 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 70/142 2026-03-10T11:28:52.217 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 71/142 2026-03-10T11:28:52.217 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 72/142 2026-03-10T11:28:52.217 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 73/142 2026-03-10T11:28:52.217 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 74/142 2026-03-10T11:28:52.217 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 75/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 76/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 77/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pluggy-0.13.1-7.el9.noarch 78/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 79/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-py-1.10.0-6.el9.noarch 80/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 81/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 82/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pytest-6.2.2-7.el9.noarch 83/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 84/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 85/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 86/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 87/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 88/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 89/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 90/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 91/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 92/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 93/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 94/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 95/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 96/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 97/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 98/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 99/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 100/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 101/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 102/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 103/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 104/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 105/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 106/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 107/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 108/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 109/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 110/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 111/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 112/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 113/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 114/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 115/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 116/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 117/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 118/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 119/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 120/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 121/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 122/142 2026-03-10T11:28:52.218 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 123/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 124/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 125/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 126/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 127/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 128/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 129/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 130/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 131/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 132/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 133/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 134/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 135/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 136/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : re2-1:20211101-20.el9.x86_64 137/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 138/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 139/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 140/142 2026-03-10T11:28:52.219 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 141/142 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 142/142 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout:Upgraded: 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout:Installed: 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.321 INFO:teuthology.orchestra.run.vm09.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-iniconfig-1.1.1-7.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T11:28:52.322 INFO:teuthology.orchestra.run.vm09.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-pluggy-0.13.1-7.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-py-1.10.0-6.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-pytest-6.2.2-7.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: zip-3.0-35.el9.x86_64 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:28:52.323 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:28:52.411 DEBUG:teuthology.parallel:result is None 2026-03-10T11:28:52.411 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T11:28:53.054 DEBUG:teuthology.orchestra.run.vm06:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-10T11:28:53.073 INFO:teuthology.orchestra.run.vm06.stdout:19.2.3-678.ge911bdeb.el9 2026-03-10T11:28:53.073 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-10T11:28:53.073 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-10T11:28:53.073 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T11:28:53.688 DEBUG:teuthology.orchestra.run.vm09:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-10T11:28:53.707 INFO:teuthology.orchestra.run.vm09.stdout:19.2.3-678.ge911bdeb.el9 2026-03-10T11:28:53.707 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-10T11:28:53.707 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-10T11:28:53.708 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-10T11:28:53.708 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:28:53.708 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T11:28:53.733 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:28:53.733 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T11:28:53.772 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-10T11:28:53.772 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:28:53.773 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T11:28:53.797 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T11:28:53.860 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:28:53.860 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T11:28:53.886 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T11:28:53.946 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-10T11:28:53.947 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:28:53.947 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T11:28:53.968 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T11:28:54.031 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:28:54.031 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T11:28:54.053 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T11:28:54.114 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-10T11:28:54.114 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:28:54.114 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T11:28:54.137 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T11:28:54.198 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:28:54.198 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T11:28:54.221 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T11:28:54.283 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-10T11:28:54.386 INFO:tasks.cephadm:Config: {'conf': {'mgr': {'debug mgr': 20, 'debug ms': 1}, 'global': {'mon election default strategy': 1, 'ms type': 'async'}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'osd shutdown pgref assert': True}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'but it is still running', 'overall HEALTH_', '\\(OSDMAP_FLAGS\\)', '\\(PG_', '\\(OSD_', '\\(OBJECT_', '\\(POOL_APP_NOT_ENABLED\\)'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'cephadm_mode': 'cephadm-package'} 2026-03-10T11:28:54.386 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T11:28:54.386 INFO:tasks.cephadm:Cluster fsid is 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:28:54.386 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-10T11:28:54.386 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.106', 'mon.c': '[v2:192.168.123.106:3301,v1:192.168.123.106:6790]', 'mon.b': '192.168.123.109'} 2026-03-10T11:28:54.386 INFO:tasks.cephadm:First mon is mon.a on vm06 2026-03-10T11:28:54.386 INFO:tasks.cephadm:First mgr is y 2026-03-10T11:28:54.386 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-10T11:28:54.386 DEBUG:teuthology.orchestra.run.vm06:> sudo hostname $(hostname -s) 2026-03-10T11:28:54.409 DEBUG:teuthology.orchestra.run.vm09:> sudo hostname $(hostname -s) 2026-03-10T11:28:54.436 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-10T11:28:54.436 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T11:28:54.451 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T11:28:54.673 INFO:teuthology.orchestra.run.vm06.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T11:28:54.677 INFO:teuthology.orchestra.run.vm09.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T11:29:37.440 INFO:teuthology.orchestra.run.vm09.stdout:{ 2026-03-10T11:29:37.440 INFO:teuthology.orchestra.run.vm09.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T11:29:37.440 INFO:teuthology.orchestra.run.vm09.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T11:29:37.440 INFO:teuthology.orchestra.run.vm09.stdout: "repo_digests": [ 2026-03-10T11:29:37.440 INFO:teuthology.orchestra.run.vm09.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T11:29:37.440 INFO:teuthology.orchestra.run.vm09.stdout: ] 2026-03-10T11:29:37.440 INFO:teuthology.orchestra.run.vm09.stdout:} 2026-03-10T11:29:43.964 INFO:teuthology.orchestra.run.vm06.stdout:{ 2026-03-10T11:29:43.964 INFO:teuthology.orchestra.run.vm06.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T11:29:43.964 INFO:teuthology.orchestra.run.vm06.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T11:29:43.964 INFO:teuthology.orchestra.run.vm06.stdout: "repo_digests": [ 2026-03-10T11:29:43.964 INFO:teuthology.orchestra.run.vm06.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T11:29:43.964 INFO:teuthology.orchestra.run.vm06.stdout: ] 2026-03-10T11:29:43.964 INFO:teuthology.orchestra.run.vm06.stdout:} 2026-03-10T11:29:43.979 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /etc/ceph 2026-03-10T11:29:44.005 DEBUG:teuthology.orchestra.run.vm09:> sudo mkdir -p /etc/ceph 2026-03-10T11:29:44.028 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 777 /etc/ceph 2026-03-10T11:29:44.068 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 777 /etc/ceph 2026-03-10T11:29:44.130 INFO:tasks.cephadm:Writing seed config... 2026-03-10T11:29:44.131 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-10T11:29:44.131 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-10T11:29:44.131 INFO:tasks.cephadm: override: [global] mon election default strategy = 1 2026-03-10T11:29:44.131 INFO:tasks.cephadm: override: [global] ms type = async 2026-03-10T11:29:44.131 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-10T11:29:44.131 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-10T11:29:44.131 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-10T11:29:44.131 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-10T11:29:44.131 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-10T11:29:44.131 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-10T11:29:44.131 INFO:tasks.cephadm: override: [osd] osd shutdown pgref assert = True 2026-03-10T11:29:44.131 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:29:44.131 DEBUG:teuthology.orchestra.run.vm06:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-10T11:29:44.145 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 52119272-1c74-11f1-990e-e7cc1c0e3258 mon election default strategy = 1 ms type = async [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = True bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-10T11:29:44.145 DEBUG:teuthology.orchestra.run.vm06:mon.a> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.a.service 2026-03-10T11:29:44.187 DEBUG:teuthology.orchestra.run.vm06:mgr.y> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mgr.y.service 2026-03-10T11:29:44.229 INFO:tasks.cephadm:Bootstrapping... 2026-03-10T11:29:44.229 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id y --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.106 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-10T11:29:44.368 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-10T11:29:44.368 INFO:teuthology.orchestra.run.vm06.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', '52119272-1c74-11f1-990e-e7cc1c0e3258', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'y', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.106', '--skip-admin-label'] 2026-03-10T11:29:44.368 INFO:teuthology.orchestra.run.vm06.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-10T11:29:44.368 INFO:teuthology.orchestra.run.vm06.stdout:Verifying podman|docker is present... 2026-03-10T11:29:44.386 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stdout 5.8.0 2026-03-10T11:29:44.386 INFO:teuthology.orchestra.run.vm06.stdout:Verifying lvm2 is present... 2026-03-10T11:29:44.386 INFO:teuthology.orchestra.run.vm06.stdout:Verifying time synchronization is in place... 2026-03-10T11:29:44.392 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T11:29:44.392 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T11:29:44.398 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T11:29:44.398 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-10T11:29:44.404 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout enabled 2026-03-10T11:29:44.409 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout active 2026-03-10T11:29:44.409 INFO:teuthology.orchestra.run.vm06.stdout:Unit chronyd.service is enabled and running 2026-03-10T11:29:44.409 INFO:teuthology.orchestra.run.vm06.stdout:Repeating the final host check... 2026-03-10T11:29:44.426 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stdout 5.8.0 2026-03-10T11:29:44.426 INFO:teuthology.orchestra.run.vm06.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-10T11:29:44.426 INFO:teuthology.orchestra.run.vm06.stdout:systemctl is present 2026-03-10T11:29:44.426 INFO:teuthology.orchestra.run.vm06.stdout:lvcreate is present 2026-03-10T11:29:44.432 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T11:29:44.432 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T11:29:44.437 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T11:29:44.437 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-10T11:29:44.442 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout enabled 2026-03-10T11:29:44.447 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout active 2026-03-10T11:29:44.447 INFO:teuthology.orchestra.run.vm06.stdout:Unit chronyd.service is enabled and running 2026-03-10T11:29:44.447 INFO:teuthology.orchestra.run.vm06.stdout:Host looks OK 2026-03-10T11:29:44.447 INFO:teuthology.orchestra.run.vm06.stdout:Cluster fsid: 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:29:44.447 INFO:teuthology.orchestra.run.vm06.stdout:Acquiring lock 139729501831952 on /run/cephadm/52119272-1c74-11f1-990e-e7cc1c0e3258.lock 2026-03-10T11:29:44.448 INFO:teuthology.orchestra.run.vm06.stdout:Lock 139729501831952 acquired on /run/cephadm/52119272-1c74-11f1-990e-e7cc1c0e3258.lock 2026-03-10T11:29:44.448 INFO:teuthology.orchestra.run.vm06.stdout:Verifying IP 192.168.123.106 port 3300 ... 2026-03-10T11:29:44.448 INFO:teuthology.orchestra.run.vm06.stdout:Verifying IP 192.168.123.106 port 6789 ... 2026-03-10T11:29:44.448 INFO:teuthology.orchestra.run.vm06.stdout:Base mon IP(s) is [192.168.123.106:3300, 192.168.123.106:6789], mon addrv is [v2:192.168.123.106:3300,v1:192.168.123.106:6789] 2026-03-10T11:29:44.451 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.106 metric 100 2026-03-10T11:29:44.451 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.106 metric 100 2026-03-10T11:29:44.453 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-10T11:29:44.453 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-10T11:29:44.455 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-10T11:29:44.455 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-10T11:29:44.455 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T11:29:44.455 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-10T11:29:44.455 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:6/64 scope link noprefixroute 2026-03-10T11:29:44.455 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T11:29:44.455 INFO:teuthology.orchestra.run.vm06.stdout:Mon IP `192.168.123.106` is in CIDR network `192.168.123.0/24` 2026-03-10T11:29:44.455 INFO:teuthology.orchestra.run.vm06.stdout:Mon IP `192.168.123.106` is in CIDR network `192.168.123.0/24` 2026-03-10T11:29:44.455 INFO:teuthology.orchestra.run.vm06.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-10T11:29:44.456 INFO:teuthology.orchestra.run.vm06.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-10T11:29:44.456 INFO:teuthology.orchestra.run.vm06.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T11:29:45.665 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stdout 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-10T11:29:45.665 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Trying to pull quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T11:29:45.665 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Getting image source signatures 2026-03-10T11:29:45.665 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Copying blob sha256:1752b8d01aa0dd33bbe0ab24e8316174c94fbdcd5d26252e2680bba0624747a7 2026-03-10T11:29:45.665 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Copying blob sha256:8e380faede39ebd4286247457b408d979ab568aafd8389c42ec304b8cfba4e92 2026-03-10T11:29:45.665 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Copying config sha256:654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-10T11:29:45.665 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-10T11:29:45.816 INFO:teuthology.orchestra.run.vm06.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T11:29:45.816 INFO:teuthology.orchestra.run.vm06.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T11:29:45.816 INFO:teuthology.orchestra.run.vm06.stdout:Extracting ceph user uid/gid from container image... 2026-03-10T11:29:45.909 INFO:teuthology.orchestra.run.vm06.stdout:stat: stdout 167 167 2026-03-10T11:29:45.909 INFO:teuthology.orchestra.run.vm06.stdout:Creating initial keys... 2026-03-10T11:29:46.004 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-authtool: stdout AQCpALBpGJkVOhAA8uuGKe2bm2t9yUy0N6dDmg== 2026-03-10T11:29:46.089 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-authtool: stdout AQCqALBpFHpmBBAAlHflZeImZ5XCA5Kc4GkPRQ== 2026-03-10T11:29:46.222 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-authtool: stdout AQCqALBp9SPRChAApfA5UXxzm1BV8l+rVjxgOw== 2026-03-10T11:29:46.223 INFO:teuthology.orchestra.run.vm06.stdout:Creating initial monmap... 2026-03-10T11:29:46.315 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T11:29:46.315 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-10T11:29:46.315 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:29:46.315 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T11:29:46.315 INFO:teuthology.orchestra.run.vm06.stdout:monmaptool for a [v2:192.168.123.106:3300,v1:192.168.123.106:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T11:29:46.315 INFO:teuthology.orchestra.run.vm06.stdout:setting min_mon_release = quincy 2026-03-10T11:29:46.315 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: set fsid to 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:29:46.315 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T11:29:46.315 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:29:46.315 INFO:teuthology.orchestra.run.vm06.stdout:Creating mon... 2026-03-10T11:29:46.477 INFO:teuthology.orchestra.run.vm06.stdout:create mon.a on 2026-03-10T11:29:46.623 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-10T11:29:46.740 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-10T11:29:46.860 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-52119272-1c74-11f1-990e-e7cc1c0e3258.target → /etc/systemd/system/ceph-52119272-1c74-11f1-990e-e7cc1c0e3258.target. 2026-03-10T11:29:46.860 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-52119272-1c74-11f1-990e-e7cc1c0e3258.target → /etc/systemd/system/ceph-52119272-1c74-11f1-990e-e7cc1c0e3258.target. 2026-03-10T11:29:47.047 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.a 2026-03-10T11:29:47.047 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to reset failed state of unit ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.a.service: Unit ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.a.service not loaded. 2026-03-10T11:29:47.180 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-52119272-1c74-11f1-990e-e7cc1c0e3258.target.wants/ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.a.service → /etc/systemd/system/ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@.service. 2026-03-10T11:29:47.366 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-10T11:29:47.367 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T11:29:47.367 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mon to start... 2026-03-10T11:29:47.367 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mon... 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout cluster: 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout id: 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout services: 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.146849s) 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout data: 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-10T11:29:47.568 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-10T11:29:47.569 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout pgs: 2026-03-10T11:29:47.569 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:47.569 INFO:teuthology.orchestra.run.vm06.stdout:mon is available 2026-03-10T11:29:47.569 INFO:teuthology.orchestra.run.vm06.stdout:Assimilating anything we can from ceph.conf... 2026-03-10T11:29:47.780 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:47.780 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T11:29:47.780 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout fsid = 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.106:3300,v1:192.168.123.106:6789] 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T11:29:47.781 INFO:teuthology.orchestra.run.vm06.stdout:Generating new minimal ceph.conf... 2026-03-10T11:29:47.959 INFO:teuthology.orchestra.run.vm06.stdout:Restarting the monitor... 2026-03-10T11:29:48.104 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a[48941]: 2026-03-10T11:29:48.040+0000 7fc22fa47640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-10T11:29:48.104 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 podman[49143]: 2026-03-10 11:29:48.104470571 +0000 UTC m=+0.076120437 container died e2c3ed26158761a382efe5ee0ca8e3d918add84dc563fa731dfa5120686c4cc6 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a, CEPH_REF=squid, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2) 2026-03-10T11:29:48.304 INFO:teuthology.orchestra.run.vm06.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 podman[49143]: 2026-03-10 11:29:48.120718925 +0000 UTC m=+0.092368782 container remove e2c3ed26158761a382efe5ee0ca8e3d918add84dc563fa731dfa5120686c4cc6 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, io.buildah.version=1.41.3, ceph=True) 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 bash[49143]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.a.service: Deactivated successfully. 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 systemd[1]: Stopped Ceph mon.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 systemd[1]: Starting Ceph mon.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 podman[49214]: 2026-03-10 11:29:48.262498174 +0000 UTC m=+0.015200514 container create b634a0f448ca74bd67da764bcfa26988d3a5ef1bf9873f6bdf0764d112c1128a (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a, ceph=True, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0) 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 podman[49214]: 2026-03-10 11:29:48.293831315 +0000 UTC m=+0.046533655 container init b634a0f448ca74bd67da764bcfa26988d3a5ef1bf9873f6bdf0764d112c1128a (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 podman[49214]: 2026-03-10 11:29:48.296441992 +0000 UTC m=+0.049144332 container start b634a0f448ca74bd67da764bcfa26988d3a5ef1bf9873f6bdf0764d112c1128a (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a, org.label-schema.build-date=20260223, OSD_FLAVOR=default, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, CEPH_REF=squid, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 bash[49214]: b634a0f448ca74bd67da764bcfa26988d3a5ef1bf9873f6bdf0764d112c1128a 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 podman[49214]: 2026-03-10 11:29:48.256346359 +0000 UTC m=+0.009048708 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 systemd[1]: Started Ceph mon.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: set uid:gid to 167:167 (ceph:ceph) 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 2 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: pidfile_write: ignore empty --pid-file 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: load: jerasure load: lrc 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: RocksDB version: 7.9.2 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Git sha 0 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: DB SUMMARY 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: DB Session ID: OCBKGKPBSR522XPB35YR 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: CURRENT file: CURRENT 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: IDENTITY file: IDENTITY 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 75933 ; 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.error_if_exists: 0 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.create_if_missing: 0 2026-03-10T11:29:48.367 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.paranoid_checks: 1 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.env: 0x555cf71b2dc0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.fs: PosixFileSystem 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.info_log: 0x555cf96e4700 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_file_opening_threads: 16 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.statistics: (nil) 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.use_fsync: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_log_file_size: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.keep_log_file_num: 1000 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.recycle_log_file_num: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.allow_fallocate: 1 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.allow_mmap_reads: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.allow_mmap_writes: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.use_direct_reads: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.create_missing_column_families: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.db_log_dir: 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.wal_dir: 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.advise_random_on_open: 1 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.db_write_buffer_size: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.write_buffer_manager: 0x555cf96e9900 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.rate_limiter: (nil) 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.wal_recovery_mode: 2 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.enable_thread_tracking: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.enable_pipelined_write: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.unordered_write: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.row_cache: None 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.wal_filter: None 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.allow_ingest_behind: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.two_write_queues: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.manual_wal_flush: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.wal_compression: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.atomic_flush: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.log_readahead_size: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.best_efforts_recovery: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.allow_data_in_errors: 0 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.db_host_id: __hostname__ 2026-03-10T11:29:48.368 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_background_jobs: 2 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_background_compactions: -1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_subcompactions: 1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_total_wal_size: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_open_files: -1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bytes_per_sync: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_readahead_size: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_background_flushes: -1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Compression algorithms supported: 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: kZSTD supported: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: kXpressCompression supported: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: kBZip2Compression supported: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: kLZ4Compression supported: 1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: kZlibCompression supported: 1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: kLZ4HCCompression supported: 1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: kSnappyCompression supported: 1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.merge_operator: 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_filter: None 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_filter_factory: None 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.sst_partitioner_factory: None 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x555cf96e4640) 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: cache_index_and_filter_blocks: 1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: pin_top_level_index_and_filter: 1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: index_type: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: data_block_index_type: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: index_shortening: 1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: checksum: 4 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: no_block_cache: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: block_cache: 0x555cf9709350 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: block_cache_name: BinnedLRUCache 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: block_cache_options: 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: capacity : 536870912 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: num_shard_bits : 4 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: strict_capacity_limit : 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: high_pri_pool_ratio: 0.000 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: block_cache_compressed: (nil) 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: persistent_cache: (nil) 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: block_size: 4096 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: block_size_deviation: 10 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: block_restart_interval: 16 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: index_block_restart_interval: 1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: metadata_block_size: 4096 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: partition_filters: 0 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: use_delta_encoding: 1 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: filter_policy: bloomfilter 2026-03-10T11:29:48.369 INFO:journalctl@ceph.mon.a.vm06.stdout: whole_key_filtering: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout: verify_compression: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout: read_amp_bytes_per_bit: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout: format_version: 5 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout: enable_index_compression: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout: block_align: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout: max_auto_readahead_size: 262144 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout: prepopulate_block_cache: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout: initial_auto_readahead_size: 8192 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout: num_file_reads_for_auto_readahead: 2 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.write_buffer_size: 33554432 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_write_buffer_number: 2 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compression: NoCompression 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bottommost_compression: Disabled 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.prefix_extractor: nullptr 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.num_levels: 7 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compression_opts.level: 32767 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compression_opts.strategy: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compression_opts.enabled: false 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.target_file_size_base: 67108864 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.arena_block_size: 1048576 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.disable_auto_compactions: 0 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T11:29:48.370 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.inplace_update_support: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.bloom_locality: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.max_successive_merges: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.paranoid_file_checks: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.force_consistency_checks: 1 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.report_bg_io_stats: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.ttl: 2592000 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.enable_blob_files: false 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.min_blob_size: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.blob_file_size: 268435456 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.blob_file_starting_level: 0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 0158ec10-f2de-4897-b2b8-36cd1ec1e4a4 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773142188318044, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773142188322289, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72911, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 227, "table_properties": {"data_size": 71190, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9804, "raw_average_key_size": 49, "raw_value_size": 65683, "raw_average_value_size": 331, "num_data_blocks": 8, "num_entries": 198, "num_filter_entries": 198, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773142188, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "0158ec10-f2de-4897-b2b8-36cd1ec1e4a4", "db_session_id": "OCBKGKPBSR522XPB35YR", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773142188322345, "job": 1, "event": "recovery_finished"} 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x555cf970ae00 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: DB pointer 0x555cf9820000 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: ** DB Stats ** 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: ** Compaction Stats [default] ** 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: L0 2/0 73.06 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 17.2 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Sum 2/0 73.06 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 17.2 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 17.2 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: ** Compaction Stats [default] ** 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 17.2 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T11:29:48.371 INFO:journalctl@ceph.mon.a.vm06.stdout: 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: Cumulative compaction: 0.00 GB write, 9.06 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: Interval compaction: 0.00 GB write, 9.06 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: Block cache BinnedLRUCache@0x555cf9709350#2 capacity: 512.00 MB usage: 1.06 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1e-05 secs_since: 0 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: Block cache entry stats(count,size,portion): FilterBlock(2,0.70 KB,0.00013411%) IndexBlock(2,0.36 KB,6.85453e-05%) Misc(1,0.00 KB,0%) 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: starting mon.a rank 0 at public addrs [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] at bind addrs [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: mon.a@-1(???) e1 preinit fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: mon.a@-1(???).mds e1 new map 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: mon.a@-1(???).mds e1 print_map 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: e1 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: btime 2026-03-10T11:29:47:393668+0000 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: legacy client fscid: -1 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout: No filesystems configured 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: mon.a@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T11:29:48.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: mon.a@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-10T11:29:48.492 INFO:teuthology.orchestra.run.vm06.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-10T11:29:48.493 INFO:teuthology.orchestra.run.vm06.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-10T11:29:48.493 INFO:teuthology.orchestra.run.vm06.stdout:Creating mgr... 2026-03-10T11:29:48.493 INFO:teuthology.orchestra.run.vm06.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-10T11:29:48.494 INFO:teuthology.orchestra.run.vm06.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-10T11:29:48.632 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T11:29:48.633 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: monmap epoch 1 2026-03-10T11:29:48.633 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:29:48.633 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: last_changed 2026-03-10T11:29:46.296870+0000 2026-03-10T11:29:48.633 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:29:48.633 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: min_mon_release 19 (squid) 2026-03-10T11:29:48.633 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: election_strategy: 1 2026-03-10T11:29:48.633 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:29:48.633 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: fsmap 2026-03-10T11:29:48.633 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: osdmap e1: 0 total, 0 up, 0 in 2026-03-10T11:29:48.633 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:48 vm06 ceph-mon[49228]: mgrmap e1: no daemons active 2026-03-10T11:29:48.652 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mgr.y 2026-03-10T11:29:48.652 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to reset failed state of unit ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mgr.y.service: Unit ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mgr.y.service not loaded. 2026-03-10T11:29:48.774 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-52119272-1c74-11f1-990e-e7cc1c0e3258.target.wants/ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mgr.y.service → /etc/systemd/system/ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@.service. 2026-03-10T11:29:48.886 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:48 vm06 systemd[1]: Starting Ceph mgr.y for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:29:48.927 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-10T11:29:48.927 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T11:29:48.927 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-10T11:29:48.927 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-10T11:29:48.927 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr to start... 2026-03-10T11:29:48.927 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr... 2026-03-10T11:29:49.139 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:48 vm06 podman[49425]: 2026-03-10 11:29:48.886249493 +0000 UTC m=+0.015484777 container create 6328e6419ad4ba02651113e4d1135a9eb1e068f78beb8db614c2ca4e3ca1f195 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-10T11:29:49.139 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:48 vm06 podman[49425]: 2026-03-10 11:29:48.91520919 +0000 UTC m=+0.044444474 container init 6328e6419ad4ba02651113e4d1135a9eb1e068f78beb8db614c2ca4e3ca1f195 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-10T11:29:49.139 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:48 vm06 podman[49425]: 2026-03-10 11:29:48.918208305 +0000 UTC m=+0.047443578 container start 6328e6419ad4ba02651113e4d1135a9eb1e068f78beb8db614c2ca4e3ca1f195 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, ceph=True, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, org.label-schema.schema-version=1.0) 2026-03-10T11:29:49.139 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:48 vm06 bash[49425]: 6328e6419ad4ba02651113e4d1135a9eb1e068f78beb8db614c2ca4e3ca1f195 2026-03-10T11:29:49.139 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:48 vm06 podman[49425]: 2026-03-10 11:29:48.880215147 +0000 UTC m=+0.009450432 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T11:29:49.139 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:48 vm06 systemd[1]: Started Ceph mgr.y for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:29:49.139 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:49.017+0000 7f331235d140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T11:29:49.139 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:49.061+0000 7f331235d140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T11:29:49.144 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:49.144 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T11:29:49.144 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsid": "52119272-1c74-11f1-990e-e7cc1c0e3258", 2026-03-10T11:29:49.144 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 0 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T11:29:47:393668+0000", 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T11:29:47.394296+0000", 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T11:29:49.145 INFO:teuthology.orchestra.run.vm06.stdout:mgr not available, waiting (1/15)... 2026-03-10T11:29:49.777 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:49 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1670769261' entity='client.admin' 2026-03-10T11:29:49.777 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:49 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2148088620' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T11:29:49.777 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:49.470+0000 7f331235d140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T11:29:49.778 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:49.776+0000 7f331235d140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T11:29:50.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T11:29:50.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T11:29:50.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: from numpy import show_config as show_numpy_config 2026-03-10T11:29:50.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:49.861+0000 7f331235d140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T11:29:50.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:49.897+0000 7f331235d140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T11:29:50.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:49.964+0000 7f331235d140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T11:29:50.693 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:50.442+0000 7f331235d140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T11:29:50.693 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:50.547+0000 7f331235d140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:29:50.693 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:50.584+0000 7f331235d140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T11:29:50.693 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:50.618+0000 7f331235d140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T11:29:50.694 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:50.657+0000 7f331235d140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T11:29:50.694 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:50.692+0000 7f331235d140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T11:29:51.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:50.857+0000 7f331235d140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T11:29:51.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:50.906+0000 7f331235d140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsid": "52119272-1c74-11f1-990e-e7cc1c0e3258", 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 0 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_age": 2, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T11:29:51.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T11:29:47:393668+0000", 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T11:29:47.394296+0000", 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T11:29:51.346 INFO:teuthology.orchestra.run.vm06.stdout:mgr not available, waiting (2/15)... 2026-03-10T11:29:51.422 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:51 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1932212887' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T11:29:51.422 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:51 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:51.119+0000 7f331235d140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T11:29:51.422 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:51 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:51.421+0000 7f331235d140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T11:29:51.682 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:51 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:51.457+0000 7f331235d140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T11:29:51.683 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:51 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:51.497+0000 7f331235d140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T11:29:51.683 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:51 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:51.571+0000 7f331235d140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T11:29:51.683 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:51 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:51.605+0000 7f331235d140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T11:29:51.683 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:51 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:51.682+0000 7f331235d140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T11:29:51.959 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:51 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:51.792+0000 7f331235d140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:29:51.959 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:51 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:51.923+0000 7f331235d140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T11:29:51.959 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:51 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:51.958+0000 7f331235d140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: Activating manager daemon y 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: mgrmap e2: y(active, starting, since 0.00365867s) 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: Manager daemon y is now available 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' 2026-03-10T11:29:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:52 vm06 ceph-mon[49228]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsid": "52119272-1c74-11f1-990e-e7cc1c0e3258", 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 0 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T11:29:53.622 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T11:29:53.623 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T11:29:47:393668+0000", 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T11:29:53.624 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T11:29:53.625 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T11:29:47.394296+0000", 2026-03-10T11:29:53.625 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T11:29:53.625 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T11:29:53.625 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T11:29:53.625 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T11:29:53.625 INFO:teuthology.orchestra.run.vm06.stdout:mgr is available 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout fsid = 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.106:3300,v1:192.168.123.106:6789] 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T11:29:53.871 INFO:teuthology.orchestra.run.vm06.stdout:Enabling cephadm module... 2026-03-10T11:29:54.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:53 vm06 ceph-mon[49228]: mgrmap e3: y(active, since 1.00798s) 2026-03-10T11:29:54.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:53 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1082594077' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T11:29:54.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:53 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3945623780' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T11:29:55.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:54 vm06 ceph-mon[49228]: mgrmap e4: y(active, since 2s) 2026-03-10T11:29:55.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:54 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/292608183' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T11:29:55.247 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ignoring --setuser ceph since I am not root 2026-03-10T11:29:55.247 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ignoring --setgroup ceph since I am not root 2026-03-10T11:29:55.247 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:55.112+0000 7f4f8dedb140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T11:29:55.247 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:55.158+0000 7f4f8dedb140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T11:29:55.277 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T11:29:55.277 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-10T11:29:55.277 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T11:29:55.277 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "active_name": "y", 2026-03-10T11:29:55.277 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T11:29:55.277 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T11:29:55.277 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for the mgr to restart... 2026-03-10T11:29:55.277 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr epoch 5... 2026-03-10T11:29:55.861 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:55.559+0000 7f4f8dedb140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T11:29:56.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:55 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/292608183' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T11:29:56.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:55 vm06 ceph-mon[49228]: mgrmap e5: y(active, since 3s) 2026-03-10T11:29:56.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:55 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2116504606' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T11:29:56.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:55.860+0000 7f4f8dedb140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T11:29:56.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T11:29:56.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T11:29:56.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: from numpy import show_config as show_numpy_config 2026-03-10T11:29:56.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:55.943+0000 7f4f8dedb140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T11:29:56.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:55.980+0000 7f4f8dedb140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T11:29:56.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:56 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:56.047+0000 7f4f8dedb140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T11:29:56.772 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:56 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:56.521+0000 7f4f8dedb140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T11:29:56.772 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:56 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:56.626+0000 7f4f8dedb140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:29:56.772 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:56 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:56.664+0000 7f4f8dedb140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T11:29:56.772 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:56 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:56.696+0000 7f4f8dedb140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T11:29:56.772 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:56 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:56.736+0000 7f4f8dedb140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T11:29:56.772 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:56 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:56.771+0000 7f4f8dedb140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T11:29:57.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:56 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:56.932+0000 7f4f8dedb140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T11:29:57.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:56 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:56.981+0000 7f4f8dedb140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T11:29:57.455 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:57 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:57.190+0000 7f4f8dedb140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T11:29:57.717 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:57 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:57.455+0000 7f4f8dedb140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T11:29:57.718 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:57 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:57.491+0000 7f4f8dedb140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T11:29:57.718 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:57 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:57.530+0000 7f4f8dedb140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T11:29:57.718 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:57 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:57.604+0000 7f4f8dedb140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T11:29:57.718 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:57 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:57.640+0000 7f4f8dedb140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T11:29:57.718 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:57 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:57.716+0000 7f4f8dedb140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T11:29:57.991 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:57 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:57.826+0000 7f4f8dedb140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:29:57.992 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:57 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:57.957+0000 7f4f8dedb140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:57 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:29:57.991+0000 7f4f8dedb140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: Active manager daemon y restarted 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: Activating manager daemon y 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: osdmap e2: 0 total, 0 up, 0 in 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: mgrmap e6: y(active, starting, since 0.00528498s) 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: Manager daemon y is now available 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: Found migration_current of "None". Setting to last migration. 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:29:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:58 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:29:59.032 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T11:29:59.032 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-10T11:29:59.032 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T11:29:59.032 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T11:29:59.032 INFO:teuthology.orchestra.run.vm06.stdout:mgr epoch 5 is available 2026-03-10T11:29:59.032 INFO:teuthology.orchestra.run.vm06.stdout:Setting orchestrator backend to cephadm... 2026-03-10T11:29:59.579 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-10T11:29:59.579 INFO:teuthology.orchestra.run.vm06.stdout:Generating ssh key... 2026-03-10T11:29:59.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:59 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:29:59.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:59 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:29:59.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:59 vm06 ceph-mon[49228]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T11:29:59.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:59 vm06 ceph-mon[49228]: mgrmap e7: y(active, since 1.01006s) 2026-03-10T11:29:59.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:59 vm06 ceph-mon[49228]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T11:29:59.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:59 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:29:59.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:59 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:29:59.816 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:29:59 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: Generating public/private ed25519 key pair. 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: Your identification has been saved in /tmp/tmpaoheaqh2/key 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: Your public key has been saved in /tmp/tmpaoheaqh2/key.pub 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: The key fingerprint is: 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: SHA256:LKfkC7B7y57p+8G6jIZ8hGPDlhW1wqJOmYXXw9NLk88 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: The key's randomart image is: 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: +--[ED25519 256]--+ 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: | .. | 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: | o.o o . | 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: | o =.* = | 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: |. *.. +.= | 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: |o+= o.SE | 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: |oO + + + | 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: |+o= . = | 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: |...*.= o | 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: | .++%=o | 2026-03-10T11:29:59.817 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:29:59 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: +----[SHA256]-----+ 2026-03-10T11:30:00.091 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIH7pUl4mQqjJmNt5njvso69XJyvuCBV60Zagr4iwplgh ceph-52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:00.091 INFO:teuthology.orchestra.run.vm06.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-10T11:30:00.091 INFO:teuthology.orchestra.run.vm06.stdout:Adding key to root@localhost authorized_keys... 2026-03-10T11:30:00.092 INFO:teuthology.orchestra.run.vm06.stdout:Adding host vm06... 2026-03-10T11:30:00.939 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:00 vm06 ceph-mon[49228]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:00.939 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:00 vm06 ceph-mon[49228]: [10/Mar/2026:11:29:59] ENGINE Bus STARTING 2026-03-10T11:30:00.939 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:00 vm06 ceph-mon[49228]: [10/Mar/2026:11:29:59] ENGINE Serving on https://192.168.123.106:7150 2026-03-10T11:30:00.939 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:00 vm06 ceph-mon[49228]: [10/Mar/2026:11:29:59] ENGINE Client ('192.168.123.106', 53842) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T11:30:00.939 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:00 vm06 ceph-mon[49228]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:00.939 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:00 vm06 ceph-mon[49228]: [10/Mar/2026:11:29:59] ENGINE Serving on http://192.168.123.106:8765 2026-03-10T11:30:00.939 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:00 vm06 ceph-mon[49228]: [10/Mar/2026:11:29:59] ENGINE Bus STARTED 2026-03-10T11:30:00.939 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:00 vm06 ceph-mon[49228]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:00.939 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:00 vm06 ceph-mon[49228]: Generating ssh key... 2026-03-10T11:30:00.939 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:00 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:00.939 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:00 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:01.843 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Added host 'vm06' with addr '192.168.123.106' 2026-03-10T11:30:01.843 INFO:teuthology.orchestra.run.vm06.stdout:Deploying unmanaged mon service... 2026-03-10T11:30:02.107 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:01 vm06 ceph-mon[49228]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:02.107 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:01 vm06 ceph-mon[49228]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm06", "addr": "192.168.123.106", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:02.107 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:01 vm06 ceph-mon[49228]: Deploying cephadm binary to vm06 2026-03-10T11:30:02.107 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:01 vm06 ceph-mon[49228]: mgrmap e8: y(active, since 2s) 2026-03-10T11:30:02.107 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:01 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:02.107 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:01 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:02.155 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-10T11:30:02.155 INFO:teuthology.orchestra.run.vm06.stdout:Deploying unmanaged mgr service... 2026-03-10T11:30:02.404 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-10T11:30:03.217 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:03 vm06 ceph-mon[49228]: Added host vm06 2026-03-10T11:30:03.217 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:03 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:03.217 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:03 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:03.217 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:03 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1047890241' entity='client.admin' 2026-03-10T11:30:03.735 INFO:teuthology.orchestra.run.vm06.stdout:Enabling the dashboard module... 2026-03-10T11:30:04.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:04 vm06 ceph-mon[49228]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:04.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:04 vm06 ceph-mon[49228]: Saving service mon spec with placement count:5 2026-03-10T11:30:04.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:04 vm06 ceph-mon[49228]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:04.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:04 vm06 ceph-mon[49228]: Saving service mgr spec with placement count:2 2026-03-10T11:30:04.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:04 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:04.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:04 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3543230720' entity='client.admin' 2026-03-10T11:30:04.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:04 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:04.372 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:04 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/209722436' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T11:30:05.081 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:04 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ignoring --setuser ceph since I am not root 2026-03-10T11:30:05.081 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:04 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ignoring --setgroup ceph since I am not root 2026-03-10T11:30:05.081 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:05 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:05.048+0000 7f2a492b2140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T11:30:05.251 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T11:30:05.251 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-10T11:30:05.251 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T11:30:05.251 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "active_name": "y", 2026-03-10T11:30:05.251 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T11:30:05.251 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T11:30:05.251 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for the mgr to restart... 2026-03-10T11:30:05.251 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr epoch 9... 2026-03-10T11:30:05.332 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:05 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:05.332 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:05 vm06 ceph-mon[49228]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:05.332 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:05 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/209722436' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T11:30:05.332 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:05 vm06 ceph-mon[49228]: mgrmap e9: y(active, since 6s) 2026-03-10T11:30:05.332 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:05 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/4197044674' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T11:30:05.333 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:05 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:05.094+0000 7f2a492b2140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T11:30:05.781 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:05 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:05.568+0000 7f2a492b2140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T11:30:06.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:05 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:05.916+0000 7f2a492b2140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T11:30:06.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T11:30:06.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T11:30:06.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: from numpy import show_config as show_numpy_config 2026-03-10T11:30:06.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:06.014+0000 7f2a492b2140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T11:30:06.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:06.054+0000 7f2a492b2140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T11:30:06.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:06.127+0000 7f2a492b2140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T11:30:06.905 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:06.641+0000 7f2a492b2140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T11:30:06.905 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:06.749+0000 7f2a492b2140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:30:06.905 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:06.788+0000 7f2a492b2140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T11:30:06.905 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:06.822+0000 7f2a492b2140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T11:30:06.905 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:06.866+0000 7f2a492b2140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T11:30:06.905 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:06 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:06.905+0000 7f2a492b2140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T11:30:07.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:07 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:07.078+0000 7f2a492b2140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T11:30:07.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:07 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:07.125+0000 7f2a492b2140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T11:30:07.643 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:07 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:07.355+0000 7f2a492b2140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T11:30:07.643 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:07 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:07.643+0000 7f2a492b2140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T11:30:07.921 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:07 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:07.681+0000 7f2a492b2140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T11:30:07.921 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:07 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:07.723+0000 7f2a492b2140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T11:30:07.921 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:07 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:07.802+0000 7f2a492b2140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T11:30:07.921 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:07 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:07.839+0000 7f2a492b2140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T11:30:08.177 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:07 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:07.921+0000 7f2a492b2140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T11:30:08.178 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:08 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:08.039+0000 7f2a492b2140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:08 vm06 ceph-mon[49228]: Active manager daemon y restarted 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:08 vm06 ceph-mon[49228]: Activating manager daemon y 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:08 vm06 ceph-mon[49228]: osdmap e3: 0 total, 0 up, 0 in 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:08 vm06 ceph-mon[49228]: mgrmap e10: y(active, starting, since 0.0369771s) 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:08 vm06 ceph-mon[49228]: Manager daemon y is now available 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:08 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:08.178+0000 7f2a492b2140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T11:30:08.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:08 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:08.215+0000 7f2a492b2140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T11:30:09.352 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T11:30:09.352 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-10T11:30:09.352 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T11:30:09.352 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T11:30:09.352 INFO:teuthology.orchestra.run.vm06.stdout:mgr epoch 9 is available 2026-03-10T11:30:09.352 INFO:teuthology.orchestra.run.vm06.stdout:Generating a dashboard self-signed certificate... 2026-03-10T11:30:09.772 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:09 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:30:09.772 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:09 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:30:09.773 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:09 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:09.773 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:09 vm06 ceph-mon[49228]: mgrmap e11: y(active, since 1.04014s) 2026-03-10T11:30:09.851 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-10T11:30:09.851 INFO:teuthology.orchestra.run.vm06.stdout:Creating initial admin user... 2026-03-10T11:30:10.345 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$/FCcm8myRGy9T/sNBOvPzuKvjwNFFYwNby4m/DwkzUZQ5XNv3lvVm", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773142210, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-10T11:30:10.345 INFO:teuthology.orchestra.run.vm06.stdout:Fetching dashboard port number... 2026-03-10T11:30:10.593 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 8443 2026-03-10T11:30:10.593 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-10T11:30:10.593 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-10T11:30:10.594 INFO:teuthology.orchestra.run.vm06.stdout:Ceph Dashboard is now available at: 2026-03-10T11:30:10.595 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:10.595 INFO:teuthology.orchestra.run.vm06.stdout: URL: https://vm06.local:8443/ 2026-03-10T11:30:10.595 INFO:teuthology.orchestra.run.vm06.stdout: User: admin 2026-03-10T11:30:10.595 INFO:teuthology.orchestra.run.vm06.stdout: Password: sjlp57gzoc 2026-03-10T11:30:10.595 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:10.595 INFO:teuthology.orchestra.run.vm06.stdout:Saving cluster configuration to /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config directory 2026-03-10T11:30:10.864 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:10 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:10.864 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:10 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:10.864 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:10 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:10.864 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:10 vm06 ceph-mon[49228]: [10/Mar/2026:11:30:09] ENGINE Bus STARTING 2026-03-10T11:30:10.864 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:10 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:10.864 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:10 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/447545308' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: sudo /sbin/cephadm shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout:Or, if you are only running a single cluster on this host: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: sudo /sbin/cephadm shell 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: ceph telemetry on 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout:For more information see: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:10.909 INFO:teuthology.orchestra.run.vm06.stdout:Bootstrap complete. 2026-03-10T11:30:10.939 INFO:tasks.cephadm:Fetching config... 2026-03-10T11:30:10.939 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:30:10.939 DEBUG:teuthology.orchestra.run.vm06:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-10T11:30:10.961 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-10T11:30:10.961 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:30:10.961 DEBUG:teuthology.orchestra.run.vm06:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-10T11:30:11.024 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-10T11:30:11.024 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:30:11.024 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/keyring of=/dev/stdout 2026-03-10T11:30:11.095 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-10T11:30:11.095 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:30:11.095 DEBUG:teuthology.orchestra.run.vm06:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-10T11:30:11.154 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-10T11:30:11.154 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIH7pUl4mQqjJmNt5njvso69XJyvuCBV60Zagr4iwplgh ceph-52119272-1c74-11f1-990e-e7cc1c0e3258' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T11:30:11.242 INFO:teuthology.orchestra.run.vm06.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIH7pUl4mQqjJmNt5njvso69XJyvuCBV60Zagr4iwplgh ceph-52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:11.266 DEBUG:teuthology.orchestra.run.vm09:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIH7pUl4mQqjJmNt5njvso69XJyvuCBV60Zagr4iwplgh ceph-52119272-1c74-11f1-990e-e7cc1c0e3258' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T11:30:11.303 INFO:teuthology.orchestra.run.vm09.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIH7pUl4mQqjJmNt5njvso69XJyvuCBV60Zagr4iwplgh ceph-52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:11.315 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-10T11:30:11.512 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:11.853 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-10T11:30:11.853 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-10T11:30:12.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:11 vm06 ceph-mon[49228]: [10/Mar/2026:11:30:09] ENGINE Serving on https://192.168.123.106:7150 2026-03-10T11:30:12.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:11 vm06 ceph-mon[49228]: [10/Mar/2026:11:30:09] ENGINE Client ('192.168.123.106', 59650) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T11:30:12.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:11 vm06 ceph-mon[49228]: [10/Mar/2026:11:30:10] ENGINE Serving on http://192.168.123.106:8765 2026-03-10T11:30:12.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:11 vm06 ceph-mon[49228]: [10/Mar/2026:11:30:10] ENGINE Bus STARTED 2026-03-10T11:30:12.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:11 vm06 ceph-mon[49228]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:12.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:11 vm06 ceph-mon[49228]: mgrmap e12: y(active, since 2s) 2026-03-10T11:30:12.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:11 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3102202714' entity='client.admin' 2026-03-10T11:30:12.048 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:11 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3502460877' entity='client.admin' 2026-03-10T11:30:12.059 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:12.383 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm09 2026-03-10T11:30:12.383 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:30:12.383 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.conf 2026-03-10T11:30:12.399 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:30:12.400 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:30:12.456 INFO:tasks.cephadm:Adding host vm09 to orchestrator... 2026-03-10T11:30:12.456 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch host add vm09 2026-03-10T11:30:12.642 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:13 vm06 ceph-mon[49228]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:13 vm06 ceph-mon[49228]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:14.505 INFO:teuthology.orchestra.run.vm06.stdout:Added host 'vm09' with addr '192.168.123.109' 2026-03-10T11:30:14.566 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch host ls --format=json 2026-03-10T11:30:14.770 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: Deploying cephadm binary to vm09 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:14.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:15.026 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:15.026 INFO:teuthology.orchestra.run.vm06.stdout:[{"addr": "192.168.123.106", "hostname": "vm06", "labels": [], "status": ""}, {"addr": "192.168.123.109", "hostname": "vm09", "labels": [], "status": ""}] 2026-03-10T11:30:15.075 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-10T11:30:15.075 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd crush tunables default 2026-03-10T11:30:15.246 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:15.922 INFO:teuthology.orchestra.run.vm06.stderr:adjusted tunables profile to default 2026-03-10T11:30:15.981 INFO:tasks.cephadm:Adding mon.a on vm06 2026-03-10T11:30:15.982 INFO:tasks.cephadm:Adding mon.c on vm06 2026-03-10T11:30:15.982 INFO:tasks.cephadm:Adding mon.b on vm09 2026-03-10T11:30:15.982 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch apply mon '3;vm06:192.168.123.106=a;vm06:[v2:192.168.123.106:3301,v1:192.168.123.106:6790]=c;vm09:192.168.123.109=b' 2026-03-10T11:30:16.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:15 vm06 ceph-mon[49228]: Added host vm09 2026-03-10T11:30:16.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:15 vm06 ceph-mon[49228]: mgrmap e13: y(active, since 6s) 2026-03-10T11:30:16.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:15 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:16.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:15 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1492480736' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T11:30:16.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:15 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:16.236 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T11:30:16.281 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T11:30:16.556 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled mon update... 2026-03-10T11:30:16.605 DEBUG:teuthology.orchestra.run.vm06:mon.c> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.c.service 2026-03-10T11:30:16.607 DEBUG:teuthology.orchestra.run.vm09:mon.b> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.b.service 2026-03-10T11:30:16.611 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T11:30:16.611 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph mon dump -f json 2026-03-10T11:30:16.837 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T11:30:16.888 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T11:30:16.923 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:16 vm06 ceph-mon[49228]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T11:30:16.923 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:16 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1492480736' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T11:30:16.923 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:16 vm06 ceph-mon[49228]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:16.923 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:16 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:17.215 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:30:17.215 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"52119272-1c74-11f1-990e-e7cc1c0e3258","modified":"2026-03-10T11:29:46.296870Z","created":"2026-03-10T11:29:46.296870Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T11:30:17.215 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm06:192.168.123.106=a;vm06:[v2:192.168.123.106:3301,v1:192.168.123.106:6790]=c;vm09:192.168.123.109=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: Saving service mon spec with placement vm06:192.168.123.106=a;vm06:[v2:192.168.123.106:3301,v1:192.168.123.106:6790]=c;vm09:192.168.123.109=b;count:3 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/1056344587' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:18.298 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T11:30:18.298 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph mon dump -f json 2026-03-10T11:30:18.634 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:30:19.059 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:30:19.059 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"52119272-1c74-11f1-990e-e7cc1c0e3258","modified":"2026-03-10T11:29:46.296870Z","created":"2026-03-10T11:29:46.296870Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T11:30:19.059 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T11:30:19.242 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mon.b@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T11:30:19.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:18 vm06 ceph-mon[49228]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:30:19.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:18 vm06 ceph-mon[49228]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:19.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:18 vm06 ceph-mon[49228]: Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:30:19.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:18 vm06 ceph-mon[49228]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:30:19.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:18 vm06 ceph-mon[49228]: Deploying daemon mon.b on vm09 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mkfs 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: monmap epoch 1 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: last_changed 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: min_mon_release 19 (squid) 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: election_strategy: 1 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: fsmap 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: osdmap e1: 0 total, 0 up, 0 in 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e1: no daemons active 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1670769261' entity='client.admin' 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2148088620' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1932212887' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Activating manager daemon y 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e2: y(active, starting, since 0.00365867s) 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:30:19.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Manager daemon y is now available 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e3: y(active, since 1.00798s) 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1082594077' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3945623780' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e4: y(active, since 2s) 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/292608183' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/292608183' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e5: y(active, since 3s) 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2116504606' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Active manager daemon y restarted 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Activating manager daemon y 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: osdmap e2: 0 total, 0 up, 0 in 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e6: y(active, starting, since 0.00528498s) 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Manager daemon y is now available 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Found migration_current of "None". Setting to last migration. 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e7: y(active, since 1.01006s) 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:19.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: [10/Mar/2026:11:29:59] ENGINE Bus STARTING 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: [10/Mar/2026:11:29:59] ENGINE Serving on https://192.168.123.106:7150 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: [10/Mar/2026:11:29:59] ENGINE Client ('192.168.123.106', 53842) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: [10/Mar/2026:11:29:59] ENGINE Serving on http://192.168.123.106:8765 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: [10/Mar/2026:11:29:59] ENGINE Bus STARTED 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Generating ssh key... 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm06", "addr": "192.168.123.106", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Deploying cephadm binary to vm06 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e8: y(active, since 2s) 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Added host vm06 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1047890241' entity='client.admin' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Saving service mon spec with placement count:5 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Saving service mgr spec with placement count:2 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3543230720' entity='client.admin' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/209722436' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/209722436' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e9: y(active, since 6s) 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/4197044674' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Active manager daemon y restarted 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Activating manager daemon y 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: osdmap e3: 0 total, 0 up, 0 in 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e10: y(active, starting, since 0.0369771s) 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Manager daemon y is now available 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e11: y(active, since 1.04014s) 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: [10/Mar/2026:11:30:09] ENGINE Bus STARTING 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/447545308' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: [10/Mar/2026:11:30:09] ENGINE Serving on https://192.168.123.106:7150 2026-03-10T11:30:19.732 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: [10/Mar/2026:11:30:09] ENGINE Client ('192.168.123.106', 59650) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: [10/Mar/2026:11:30:10] ENGINE Serving on http://192.168.123.106:8765 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: [10/Mar/2026:11:30:10] ENGINE Bus STARTED 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e12: y(active, since 2s) 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3102202714' entity='client.admin' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3502460877' entity='client.admin' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Deploying cephadm binary to vm09 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Added host vm09 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mgrmap e13: y(active, since 6s) 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1492480736' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1492480736' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm06:192.168.123.106=a;vm06:[v2:192.168.123.106:3301,v1:192.168.123.106:6790]=c;vm09:192.168.123.109=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Saving service mon spec with placement vm06:192.168.123.106=a;vm06:[v2:192.168.123.106:3301,v1:192.168.123.106:6790]=c;vm09:192.168.123.109=b;count:3 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/1056344587' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:30:19.733 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:30:19.734 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: Deploying daemon mon.b on vm09 2026-03-10T11:30:19.734 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:19 vm09 ceph-mon[54793]: mon.b@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-10T11:30:20.062 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:19 vm06 systemd[1]: Starting Ceph mon.c for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:30:20.128 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T11:30:20.129 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph mon dump -f json 2026-03-10T11:30:20.310 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 podman[57391]: 2026-03-10 11:30:20.062542269 +0000 UTC m=+0.016025228 container create 0039930a66a483ff63b4a211e778e81bad2f5a06670fe5dd70600fa6f6ae7bcd (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-c, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, CEPH_REF=squid, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 podman[57391]: 2026-03-10 11:30:20.103613248 +0000 UTC m=+0.057096216 container init 0039930a66a483ff63b4a211e778e81bad2f5a06670fe5dd70600fa6f6ae7bcd (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-c, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 podman[57391]: 2026-03-10 11:30:20.105979248 +0000 UTC m=+0.059462216 container start 0039930a66a483ff63b4a211e778e81bad2f5a06670fe5dd70600fa6f6ae7bcd (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-c, io.buildah.version=1.41.3, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223) 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 bash[57391]: 0039930a66a483ff63b4a211e778e81bad2f5a06670fe5dd70600fa6f6ae7bcd 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 podman[57391]: 2026-03-10 11:30:20.056141519 +0000 UTC m=+0.009624496 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 systemd[1]: Started Ceph mon.c for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: set uid:gid to 167:167 (ceph:ceph) 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 2 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: pidfile_write: ignore empty --pid-file 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: load: jerasure load: lrc 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: RocksDB version: 7.9.2 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Git sha 0 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: DB SUMMARY 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: DB Session ID: B5WJ1AF325C7A204DVQ1 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: CURRENT file: CURRENT 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: IDENTITY file: IDENTITY 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: SST files in /var/lib/ceph/mon/ceph-c/store.db dir, Total Num: 0, files: 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-c/store.db: 000004.log size: 511 ; 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.error_if_exists: 0 2026-03-10T11:30:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.create_if_missing: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.paranoid_checks: 1 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.env: 0x55e13ce48dc0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.fs: PosixFileSystem 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.info_log: 0x55e13df4e700 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_file_opening_threads: 16 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.statistics: (nil) 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.use_fsync: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_log_file_size: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.keep_log_file_num: 1000 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.recycle_log_file_num: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.allow_fallocate: 1 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.allow_mmap_reads: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.allow_mmap_writes: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.use_direct_reads: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.create_missing_column_families: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.db_log_dir: 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.wal_dir: 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.advise_random_on_open: 1 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.db_write_buffer_size: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.write_buffer_manager: 0x55e13df53900 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.rate_limiter: (nil) 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.wal_recovery_mode: 2 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.enable_thread_tracking: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.enable_pipelined_write: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.unordered_write: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.row_cache: None 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.wal_filter: None 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.allow_ingest_behind: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.two_write_queues: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.manual_wal_flush: 0 2026-03-10T11:30:20.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.wal_compression: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.atomic_flush: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.log_readahead_size: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.best_efforts_recovery: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.allow_data_in_errors: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.db_host_id: __hostname__ 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_background_jobs: 2 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_background_compactions: -1 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_subcompactions: 1 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_total_wal_size: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_open_files: -1 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bytes_per_sync: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_readahead_size: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_background_flushes: -1 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Compression algorithms supported: 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: kZSTD supported: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: kXpressCompression supported: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: kBZip2Compression supported: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: kLZ4Compression supported: 1 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: kZlibCompression supported: 1 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: kLZ4HCCompression supported: 1 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: kSnappyCompression supported: 1 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-c/store.db/MANIFEST-000005 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.merge_operator: 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_filter: None 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_filter_factory: None 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.sst_partitioner_factory: None 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55e13df4e640) 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout: cache_index_and_filter_blocks: 1 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout: pin_top_level_index_and_filter: 1 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout: index_type: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout: data_block_index_type: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout: index_shortening: 1 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout: checksum: 4 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout: no_block_cache: 0 2026-03-10T11:30:20.533 INFO:journalctl@ceph.mon.c.vm06.stdout: block_cache: 0x55e13df73350 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: block_cache_name: BinnedLRUCache 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: block_cache_options: 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: capacity : 536870912 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: num_shard_bits : 4 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: strict_capacity_limit : 0 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: high_pri_pool_ratio: 0.000 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: block_cache_compressed: (nil) 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: persistent_cache: (nil) 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: block_size: 4096 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: block_size_deviation: 10 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: block_restart_interval: 16 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: index_block_restart_interval: 1 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: metadata_block_size: 4096 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: partition_filters: 0 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: use_delta_encoding: 1 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: filter_policy: bloomfilter 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: whole_key_filtering: 1 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: verify_compression: 0 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: read_amp_bytes_per_bit: 0 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: format_version: 5 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: enable_index_compression: 1 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: block_align: 0 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: max_auto_readahead_size: 262144 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: prepopulate_block_cache: 0 2026-03-10T11:30:20.534 INFO:journalctl@ceph.mon.c.vm06.stdout: initial_auto_readahead_size: 8192 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout: num_file_reads_for_auto_readahead: 2 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.write_buffer_size: 33554432 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_write_buffer_number: 2 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compression: NoCompression 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bottommost_compression: Disabled 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.prefix_extractor: nullptr 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.num_levels: 7 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compression_opts.level: 32767 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compression_opts.strategy: 0 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compression_opts.enabled: false 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.target_file_size_base: 67108864 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.arena_block_size: 1048576 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.disable_auto_compactions: 0 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T11:30:20.535 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.inplace_update_support: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.bloom_locality: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.max_successive_merges: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.paranoid_file_checks: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.force_consistency_checks: 1 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.report_bg_io_stats: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.ttl: 2592000 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.enable_blob_files: false 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.min_blob_size: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.blob_file_size: 268435456 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.blob_file_starting_level: 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-c/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: e68d7937-e489-4d59-923b-bdee27ce4031 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773142220135445, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773142220136176, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1643, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773142220, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "e68d7937-e489-4d59-923b-bdee27ce4031", "db_session_id": "B5WJ1AF325C7A204DVQ1", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773142220136249, "job": 1, "event": "recovery_finished"} 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-c/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55e13df74e00 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: DB pointer 0x55e13e08a000 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c does not exist in monmap, will attempt to join an existing cluster 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: using public_addrv [v2:192.168.123.106:3301/0,v1:192.168.123.106:6790/0] 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: ** DB Stats ** 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: ** Compaction Stats [default] ** 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: L0 1/0 1.60 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: Sum 1/0 1.60 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T11:30:20.536 INFO:journalctl@ceph.mon.c.vm06.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: ** Compaction Stats [default] ** 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 2.3 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: Cumulative compaction: 0.00 GB write, 0.33 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: Interval compaction: 0.00 GB write, 0.33 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: Block cache BinnedLRUCache@0x55e13df73350#2 capacity: 512.00 MB usage: 0.86 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.3e-05 secs_since: 0 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: Block cache entry stats(count,size,portion): DataBlock(1,0.64 KB,0.00012219%) FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: starting mon.c rank -1 at public addrs [v2:192.168.123.106:3301/0,v1:192.168.123.106:6790/0] at bind addrs [v2:192.168.123.106:3301/0,v1:192.168.123.106:6790/0] mon_data /var/lib/ceph/mon/ceph-c fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(???) e0 preinit fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).mds e1 new map 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).mds e1 print_map 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: e1 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: btime 2026-03-10T11:29:47:393668+0000 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: legacy client fscid: -1 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout: No filesystems configured 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mkfs 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: monmap epoch 1 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: last_changed 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: min_mon_release 19 (squid) 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: election_strategy: 1 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: fsmap 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: osdmap e1: 0 total, 0 up, 0 in 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e1: no daemons active 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1670769261' entity='client.admin' 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2148088620' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1932212887' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Activating manager daemon y 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e2: y(active, starting, since 0.00365867s) 2026-03-10T11:30:20.537 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Manager daemon y is now available 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14100 192.168.123.106:0/605981051' entity='mgr.y' 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e3: y(active, since 1.00798s) 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1082594077' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3945623780' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e4: y(active, since 2s) 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/292608183' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/292608183' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e5: y(active, since 3s) 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2116504606' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Active manager daemon y restarted 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Activating manager daemon y 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: osdmap e2: 0 total, 0 up, 0 in 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e6: y(active, starting, since 0.00528498s) 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Manager daemon y is now available 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Found migration_current of "None". Setting to last migration. 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e7: y(active, since 1.01006s) 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: [10/Mar/2026:11:29:59] ENGINE Bus STARTING 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: [10/Mar/2026:11:29:59] ENGINE Serving on https://192.168.123.106:7150 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: [10/Mar/2026:11:29:59] ENGINE Client ('192.168.123.106', 53842) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: [10/Mar/2026:11:29:59] ENGINE Serving on http://192.168.123.106:8765 2026-03-10T11:30:20.538 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: [10/Mar/2026:11:29:59] ENGINE Bus STARTED 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Generating ssh key... 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm06", "addr": "192.168.123.106", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Deploying cephadm binary to vm06 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e8: y(active, since 2s) 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Added host vm06 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1047890241' entity='client.admin' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Saving service mon spec with placement count:5 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Saving service mgr spec with placement count:2 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3543230720' entity='client.admin' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/209722436' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14118 192.168.123.106:0/813148332' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/209722436' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e9: y(active, since 6s) 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/4197044674' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Active manager daemon y restarted 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Activating manager daemon y 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: osdmap e3: 0 total, 0 up, 0 in 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e10: y(active, starting, since 0.0369771s) 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Manager daemon y is now available 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e11: y(active, since 1.04014s) 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: [10/Mar/2026:11:30:09] ENGINE Bus STARTING 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/447545308' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: [10/Mar/2026:11:30:09] ENGINE Serving on https://192.168.123.106:7150 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: [10/Mar/2026:11:30:09] ENGINE Client ('192.168.123.106', 59650) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: [10/Mar/2026:11:30:10] ENGINE Serving on http://192.168.123.106:8765 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: [10/Mar/2026:11:30:10] ENGINE Bus STARTED 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e12: y(active, since 2s) 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3102202714' entity='client.admin' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3502460877' entity='client.admin' 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:20.539 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Deploying cephadm binary to vm09 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Added host vm09 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mgrmap e13: y(active, since 6s) 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1492480736' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1492480736' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm06:192.168.123.106=a;vm06:[v2:192.168.123.106:3301,v1:192.168.123.106:6790]=c;vm09:192.168.123.109=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Saving service mon spec with placement vm06:192.168.123.106=a;vm06:[v2:192.168.123.106:3301,v1:192.168.123.106:6790]=c;vm09:192.168.123.109=b;count:3 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/1056344587' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: Deploying daemon mon.b on vm09 2026-03-10T11:30:20.540 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:20 vm06 ceph-mon[57405]: mon.c@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: Deploying daemon mon.c on vm06 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: mon.a calling monitor election 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: mon.b calling monitor election 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: monmap epoch 2 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: last_changed 2026-03-10T11:30:19.254756+0000 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: min_mon_release 19 (squid) 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: election_strategy: 1 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: fsmap 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: mgrmap e13: y(active, since 16s) 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: overall HEALTH_OK 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:24.604 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: Deploying daemon mon.c on vm06 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: mon.a calling monitor election 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: mon.b calling monitor election 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: monmap epoch 2 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: last_changed 2026-03-10T11:30:19.254756+0000 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: min_mon_release 19 (squid) 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: election_strategy: 1 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-10T11:30:24.612 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: fsmap 2026-03-10T11:30:24.613 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:24.613 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: mgrmap e13: y(active, since 16s) 2026-03-10T11:30:24.613 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: overall HEALTH_OK 2026-03-10T11:30:24.613 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:24.613 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:24.613 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:24.613 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.004 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:30:25.004 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":2,"fsid":"52119272-1c74-11f1-990e-e7cc1c0e3258","modified":"2026-03-10T11:30:19.254756Z","created":"2026-03-10T11:29:46.296870Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:3300","nonce":0},{"type":"v1","addr":"192.168.123.109:6789","nonce":0}]},"addr":"192.168.123.109:6789/0","public_addr":"192.168.123.109:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-10T11:30:25.004 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 2 2026-03-10T11:30:25.640 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:25.640 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.640 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.640 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.640 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.640 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:25.640 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:25.640 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:30:25.640 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:30:25.640 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:25.640 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/1581030423' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:25.641 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:25.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/1581030423' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:25.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:26.056 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T11:30:26.056 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph mon dump -f json 2026-03-10T11:30:26.221 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:26 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:26.254+0000 7f2a15614640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: Deploying daemon mon.c on vm06 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: mon.a calling monitor election 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: mon.b calling monitor election 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: monmap epoch 2 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: last_changed 2026-03-10T11:30:19.254756+0000 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: min_mon_release 19 (squid) 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: election_strategy: 1 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: fsmap 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: mgrmap e13: y(active, since 16s) 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: overall HEALTH_OK 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:30:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/1581030423' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:26.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: mon.a calling monitor election 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: mon.b calling monitor election 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: monmap epoch 3 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: last_changed 2026-03-10T11:30:26.176389+0000 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: min_mon_release 19 (squid) 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: election_strategy: 1 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: 2: [v2:192.168.123.106:3301/0,v1:192.168.123.106:6790/0] mon.c 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: fsmap 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: mgrmap e13: y(active, since 23s) 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: overall HEALTH_OK 2026-03-10T11:30:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:31 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: mon.a calling monitor election 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: mon.b calling monitor election 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: monmap epoch 3 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: last_changed 2026-03-10T11:30:26.176389+0000 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: min_mon_release 19 (squid) 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: election_strategy: 1 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: 2: [v2:192.168.123.106:3301/0,v1:192.168.123.106:6790/0] mon.c 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: fsmap 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: mgrmap e13: y(active, since 23s) 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: overall HEALTH_OK 2026-03-10T11:30:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:31 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:32.489 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:30:32.489 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":3,"fsid":"52119272-1c74-11f1-990e-e7cc1c0e3258","modified":"2026-03-10T11:30:26.176389Z","created":"2026-03-10T11:29:46.296870Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:3300","nonce":0},{"type":"v1","addr":"192.168.123.109:6789","nonce":0}]},"addr":"192.168.123.109:6789/0","public_addr":"192.168.123.109:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3301","nonce":0},{"type":"v1","addr":"192.168.123.106:6790","nonce":0}]},"addr":"192.168.123.106:6790/0","public_addr":"192.168.123.106:6790/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-10T11:30:32.489 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 3 2026-03-10T11:30:32.532 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-10T11:30:32.532 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph config generate-minimal-conf 2026-03-10T11:30:32.537 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:32 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:32.555 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:32 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:32.700 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:32.935 INFO:teuthology.orchestra.run.vm06.stdout:# minimal ceph.conf for 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:32.935 INFO:teuthology.orchestra.run.vm06.stdout:[global] 2026-03-10T11:30:32.935 INFO:teuthology.orchestra.run.vm06.stdout: fsid = 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:32.935 INFO:teuthology.orchestra.run.vm06.stdout: mon_host = [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] [v2:192.168.123.106:3301/0,v1:192.168.123.106:6790/0] 2026-03-10T11:30:32.981 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-10T11:30:32.982 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:30:32.982 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T11:30:33.050 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:30:33.050 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:30:33.115 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:30:33.115 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T11:30:33.143 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:30:33.143 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:30:33.210 INFO:tasks.cephadm:Adding mgr.y on vm06 2026-03-10T11:30:33.210 INFO:tasks.cephadm:Adding mgr.x on vm09 2026-03-10T11:30:33.210 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch apply mgr '2;vm06=y;vm09=x' 2026-03-10T11:30:33.425 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: mon.c calling monitor election 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: mon.c calling monitor election 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: mon.b calling monitor election 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: mon.a calling monitor election 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: mon.a is new leader, mons a,b,c in quorum (ranks 0,1,2) 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: monmap epoch 3 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: last_changed 2026-03-10T11:30:26.176389+0000 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: min_mon_release 19 (squid) 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: election_strategy: 1 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: 2: [v2:192.168.123.106:3301/0,v1:192.168.123.106:6790/0] mon.c 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: fsmap 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: mgrmap e13: y(active, since 24s) 2026-03-10T11:30:33.453 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:33 vm09 ceph-mon[54793]: overall HEALTH_OK 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: mon.c calling monitor election 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: mon.c calling monitor election 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: mon.b calling monitor election 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: mon.a calling monitor election 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: mon.a is new leader, mons a,b,c in quorum (ranks 0,1,2) 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: monmap epoch 3 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: last_changed 2026-03-10T11:30:26.176389+0000 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: min_mon_release 19 (squid) 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: election_strategy: 1 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: 2: [v2:192.168.123.106:3301/0,v1:192.168.123.106:6790/0] mon.c 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: fsmap 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: mgrmap e13: y(active, since 24s) 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[49228]: overall HEALTH_OK 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: mon.a calling monitor election 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: mon.b calling monitor election 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:33.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: monmap epoch 3 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: last_changed 2026-03-10T11:30:26.176389+0000 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: min_mon_release 19 (squid) 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: election_strategy: 1 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: 2: [v2:192.168.123.106:3301/0,v1:192.168.123.106:6790/0] mon.c 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: fsmap 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: mgrmap e13: y(active, since 23s) 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: overall HEALTH_OK 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: mon.c calling monitor election 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: mon.c calling monitor election 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: mon.b calling monitor election 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: mon.a calling monitor election 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: mon.a is new leader, mons a,b,c in quorum (ranks 0,1,2) 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: monmap epoch 3 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: last_changed 2026-03-10T11:30:26.176389+0000 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: created 2026-03-10T11:29:46.296870+0000 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: min_mon_release 19 (squid) 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: election_strategy: 1 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.b 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: 2: [v2:192.168.123.106:3301/0,v1:192.168.123.106:6790/0] mon.c 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: fsmap 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: mgrmap e13: y(active, since 24s) 2026-03-10T11:30:33.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:33 vm06 ceph-mon[57405]: overall HEALTH_OK 2026-03-10T11:30:33.668 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled mgr update... 2026-03-10T11:30:33.720 DEBUG:teuthology.orchestra.run.vm09:mgr.x> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mgr.x.service 2026-03-10T11:30:33.722 INFO:tasks.cephadm:Deploying OSDs... 2026-03-10T11:30:33.722 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:30:33.722 DEBUG:teuthology.orchestra.run.vm06:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T11:30:33.740 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T11:30:33.741 DEBUG:teuthology.orchestra.run.vm06:> ls /dev/[sv]d? 2026-03-10T11:30:33.807 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vda 2026-03-10T11:30:33.807 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdb 2026-03-10T11:30:33.807 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdc 2026-03-10T11:30:33.807 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdd 2026-03-10T11:30:33.807 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vde 2026-03-10T11:30:33.807 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T11:30:33.807 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T11:30:33.807 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdb 2026-03-10T11:30:33.870 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdb 2026-03-10T11:30:33.870 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T11:30:33.870 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 221 Links: 1 Device type: fc,10 2026-03-10T11:30:33.870 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T11:30:33.870 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T11:30:33.870 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 11:30:12.918013786 +0000 2026-03-10T11:30:33.870 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 11:27:48.146834004 +0000 2026-03-10T11:30:33.870 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 11:27:48.146834004 +0000 2026-03-10T11:30:33.870 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 11:24:33.285000000 +0000 2026-03-10T11:30:33.870 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T11:30:33.949 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T11:30:33.949 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T11:30:33.949 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000153537 s, 3.3 MB/s 2026-03-10T11:30:33.950 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T11:30:34.009 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdc 2026-03-10T11:30:34.066 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdc 2026-03-10T11:30:34.066 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T11:30:34.066 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 224 Links: 1 Device type: fc,20 2026-03-10T11:30:34.066 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T11:30:34.066 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T11:30:34.066 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 11:30:12.948013820 +0000 2026-03-10T11:30:34.066 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 11:27:48.141834001 +0000 2026-03-10T11:30:34.066 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 11:27:48.141834001 +0000 2026-03-10T11:30:34.066 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 11:24:33.296000000 +0000 2026-03-10T11:30:34.066 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T11:30:34.130 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T11:30:34.130 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T11:30:34.130 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000183253 s, 2.8 MB/s 2026-03-10T11:30:34.131 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T11:30:34.191 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdd 2026-03-10T11:30:34.250 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdd 2026-03-10T11:30:34.250 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T11:30:34.250 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 225 Links: 1 Device type: fc,30 2026-03-10T11:30:34.250 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T11:30:34.250 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T11:30:34.250 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 11:30:12.984013861 +0000 2026-03-10T11:30:34.250 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 11:27:48.114833983 +0000 2026-03-10T11:30:34.250 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 11:27:48.114833983 +0000 2026-03-10T11:30:34.250 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 11:24:33.301000000 +0000 2026-03-10T11:30:34.250 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T11:30:34.314 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T11:30:34.315 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T11:30:34.315 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000195757 s, 2.6 MB/s 2026-03-10T11:30:34.315 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T11:30:34.372 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vde 2026-03-10T11:30:34.430 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vde 2026-03-10T11:30:34.431 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T11:30:34.431 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 226 Links: 1 Device type: fc,40 2026-03-10T11:30:34.431 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T11:30:34.431 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T11:30:34.431 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 11:30:13.009013890 +0000 2026-03-10T11:30:34.431 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 11:27:48.097833972 +0000 2026-03-10T11:30:34.431 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 11:27:48.097833972 +0000 2026-03-10T11:30:34.431 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 11:24:33.305000000 +0000 2026-03-10T11:30:34.431 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T11:30:34.498 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T11:30:34.498 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T11:30:34.498 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000168025 s, 3.0 MB/s 2026-03-10T11:30:34.499 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T11:30:34.559 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:30:34.559 DEBUG:teuthology.orchestra.run.vm09:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T11:30:34.576 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T11:30:34.576 DEBUG:teuthology.orchestra.run.vm09:> ls /dev/[sv]d? 2026-03-10T11:30:34.633 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vda 2026-03-10T11:30:34.633 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdb 2026-03-10T11:30:34.633 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdc 2026-03-10T11:30:34.633 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdd 2026-03-10T11:30:34.633 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vde 2026-03-10T11:30:34.633 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T11:30:34.633 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T11:30:34.633 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdb 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm06=y;vm09=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: Saving service mgr spec with placement vm06=y;vm09=x;count:2 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T11:30:34.671 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T11:30:34.672 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T11:30:34.672 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:34.672 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:34.730 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdb 2026-03-10T11:30:34.730 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T11:30:34.730 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-10T11:30:34.730 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T11:30:34.730 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T11:30:34.730 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 11:30:16.914840805 +0000 2026-03-10T11:30:34.730 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 11:27:48.336765276 +0000 2026-03-10T11:30:34.730 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 11:27:48.336765276 +0000 2026-03-10T11:30:34.730 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 11:25:03.252000000 +0000 2026-03-10T11:30:34.730 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T11:30:34.759 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T11:30:34.759 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T11:30:34.759 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.0001014 s, 5.0 MB/s 2026-03-10T11:30:34.760 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm06=y;vm09=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: Saving service mgr spec with placement vm06=y;vm09=x;count:2 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:34.803 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:34 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:34.801+0000 7f1162b66140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T11:30:34.828 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdc 2026-03-10T11:30:34.935 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdc 2026-03-10T11:30:34.935 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T11:30:34.935 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-10T11:30:34.935 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T11:30:34.935 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T11:30:34.935 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 11:30:16.979840870 +0000 2026-03-10T11:30:34.935 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 11:27:48.339765280 +0000 2026-03-10T11:30:34.935 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 11:27:48.339765280 +0000 2026-03-10T11:30:34.935 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 11:25:03.259000000 +0000 2026-03-10T11:30:34.936 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T11:30:35.004 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T11:30:35.004 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T11:30:35.004 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000133399 s, 3.8 MB/s 2026-03-10T11:30:35.005 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm06=y;vm09=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: Saving service mgr spec with placement vm06=y;vm09=x;count:2 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:35.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:30:35.078 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdd 2026-03-10T11:30:35.127 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdd 2026-03-10T11:30:35.127 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T11:30:35.127 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-10T11:30:35.127 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T11:30:35.127 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T11:30:35.127 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 11:30:17.025840917 +0000 2026-03-10T11:30:35.127 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 11:27:48.361765306 +0000 2026-03-10T11:30:35.127 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 11:27:48.361765306 +0000 2026-03-10T11:30:35.127 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 11:25:03.263000000 +0000 2026-03-10T11:30:35.127 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T11:30:35.165 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:34 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:34.849+0000 7f1162b66140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T11:30:35.178 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T11:30:35.178 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T11:30:35.178 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.00126536 s, 405 kB/s 2026-03-10T11:30:35.180 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T11:30:35.271 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vde 2026-03-10T11:30:35.300 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vde 2026-03-10T11:30:35.300 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T11:30:35.300 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T11:30:35.300 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T11:30:35.300 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T11:30:35.300 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 11:30:17.066840958 +0000 2026-03-10T11:30:35.300 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 11:27:48.345765287 +0000 2026-03-10T11:30:35.300 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 11:27:48.345765287 +0000 2026-03-10T11:30:35.300 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 11:25:03.271000000 +0000 2026-03-10T11:30:35.301 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T11:30:35.393 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T11:30:35.393 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T11:30:35.393 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.00014927 s, 3.4 MB/s 2026-03-10T11:30:35.394 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T11:30:35.418 INFO:tasks.cephadm:Deploying osd.0 on vm06 with /dev/vde... 2026-03-10T11:30:35.418 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- lvm zap /dev/vde 2026-03-10T11:30:35.441 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:30:35 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:30:35.168+0000 7f2a15614640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-10T11:30:35.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:35.306+0000 7f1162b66140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T11:30:35.598 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: Deploying daemon mgr.x on vm09 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:35.643+0000 7f1162b66140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: from numpy import show_config as show_numpy_config 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:35.730+0000 7f1162b66140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:35.769+0000 7f1162b66140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T11:30:35.980 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:35.839+0000 7f1162b66140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: Deploying daemon mgr.x on vm09 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: Deploying daemon mgr.x on vm09 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:36.440 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:36.456 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch daemon add osd vm06:/dev/vde 2026-03-10T11:30:36.577 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:36.313+0000 7f1162b66140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T11:30:36.577 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:36.420+0000 7f1162b66140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:30:36.577 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:36.459+0000 7f1162b66140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T11:30:36.577 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:36.497+0000 7f1162b66140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T11:30:36.577 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:36.537+0000 7f1162b66140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T11:30:36.577 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:36.576+0000 7f1162b66140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T11:30:36.725 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: Reconfiguring mon.a (monmap changed)... 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: Reconfiguring daemon mon.a on vm06 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: Reconfiguring mgr.y (unknown last config time)... 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.y", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: Reconfiguring daemon mgr.y on vm06 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:36.890 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: Reconfiguring mon.a (monmap changed)... 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: Reconfiguring daemon mon.a on vm06 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: Reconfiguring mgr.y (unknown last config time)... 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.y", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: Reconfiguring daemon mgr.y on vm06 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:36.891 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: Reconfiguring mon.a (monmap changed)... 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: Reconfiguring daemon mon.a on vm06 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: Reconfiguring mgr.y (unknown last config time)... 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.y", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: Reconfiguring daemon mgr.y on vm06 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:36.981 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:36.981 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:36.982 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:36.742+0000 7f1162b66140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T11:30:36.982 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:36.792+0000 7f1162b66140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T11:30:37.357 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:37.022+0000 7f1162b66140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T11:30:37.357 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:37.356+0000 7f1162b66140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T11:30:37.642 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:37.392+0000 7f1162b66140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T11:30:37.642 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:37.438+0000 7f1162b66140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T11:30:37.643 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:37.518+0000 7f1162b66140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T11:30:37.643 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:37.555+0000 7f1162b66140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T11:30:37.643 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:37.641+0000 7f1162b66140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T11:30:37.898 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:37.751+0000 7f1162b66140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:30:37.898 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:37.897+0000 7f1162b66140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: Reconfiguring mon.c (monmap changed)... 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: Reconfiguring daemon mon.c on vm06 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: Reconfiguring mon.b (monmap changed)... 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: Reconfiguring daemon mon.b on vm09 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1040570163' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7704abb8-f49c-4aef-89e1-d9e67d3e765b"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7704abb8-f49c-4aef-89e1-d9e67d3e765b"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7704abb8-f49c-4aef-89e1-d9e67d3e765b"}]': finished 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: osdmap e5: 1 total, 0 up, 1 in 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: Standby manager daemon x started 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/crt"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/key"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:37 vm09 ceph-mon[54793]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T11:30:38.230 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:30:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:30:37.933+0000 7f1162b66140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: Reconfiguring mon.c (monmap changed)... 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: Reconfiguring daemon mon.c on vm06 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: Reconfiguring mon.b (monmap changed)... 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: Reconfiguring daemon mon.b on vm09 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1040570163' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7704abb8-f49c-4aef-89e1-d9e67d3e765b"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7704abb8-f49c-4aef-89e1-d9e67d3e765b"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7704abb8-f49c-4aef-89e1-d9e67d3e765b"}]': finished 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: osdmap e5: 1 total, 0 up, 1 in 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: Standby manager daemon x started 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/crt"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/key"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[49228]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: Reconfiguring mon.c (monmap changed)... 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: Reconfiguring daemon mon.c on vm06 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: Reconfiguring mon.b (monmap changed)... 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: Reconfiguring daemon mon.b on vm09 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1040570163' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7704abb8-f49c-4aef-89e1-d9e67d3e765b"}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7704abb8-f49c-4aef-89e1-d9e67d3e765b"}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7704abb8-f49c-4aef-89e1-d9e67d3e765b"}]': finished 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: osdmap e5: 1 total, 0 up, 1 in 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: Standby manager daemon x started 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/crt"}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/key"}]: dispatch 2026-03-10T11:30:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:37 vm06 ceph-mon[57405]: from='mgr.? 192.168.123.109:0/1730482123' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T11:30:39.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:38 vm09 ceph-mon[54793]: from='client.14211 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:39.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:38 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3923828516' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:30:39.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:38 vm09 ceph-mon[54793]: mgrmap e14: y(active, since 30s), standbys: x 2026-03-10T11:30:39.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:38 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T11:30:39.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:38 vm06 ceph-mon[49228]: from='client.14211 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:38 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3923828516' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:30:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:38 vm06 ceph-mon[49228]: mgrmap e14: y(active, since 30s), standbys: x 2026-03-10T11:30:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:38 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T11:30:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:38 vm06 ceph-mon[57405]: from='client.14211 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:38 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3923828516' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:30:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:38 vm06 ceph-mon[57405]: mgrmap e14: y(active, since 30s), standbys: x 2026-03-10T11:30:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:38 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T11:30:40.215 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:39 vm06 ceph-mon[49228]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:40.215 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:39 vm06 ceph-mon[57405]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:40.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:39 vm09 ceph-mon[54793]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:41.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:40 vm06 ceph-mon[49228]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:41.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:40 vm06 ceph-mon[57405]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:41.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:40 vm09 ceph-mon[54793]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:42.186 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:41 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T11:30:42.187 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:41 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:42.187 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:41 vm06 ceph-mon[57405]: Deploying daemon osd.0 on vm06 2026-03-10T11:30:42.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:41 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T11:30:42.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:41 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:42.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:41 vm06 ceph-mon[49228]: Deploying daemon osd.0 on vm06 2026-03-10T11:30:42.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:41 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T11:30:42.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:41 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:42.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:41 vm09 ceph-mon[54793]: Deploying daemon osd.0 on vm06 2026-03-10T11:30:42.977 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:42 vm06 ceph-mon[49228]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:43.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:42 vm09 ceph-mon[54793]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:43.248 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:42 vm06 ceph-mon[57405]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:43.979 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:43 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:43.979 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:43 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:43.979 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:43 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:43.980 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:43 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:43.980 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:43 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:43.980 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:43 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:44.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:43 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:44.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:43 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:44.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:43 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:44.785 INFO:teuthology.orchestra.run.vm06.stdout:Created osd(s) 0 on host 'vm06' 2026-03-10T11:30:44.839 DEBUG:teuthology.orchestra.run.vm06:osd.0> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.0.service 2026-03-10T11:30:44.840 INFO:tasks.cephadm:Deploying osd.1 on vm06 with /dev/vdd... 2026-03-10T11:30:44.840 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- lvm zap /dev/vdd 2026-03-10T11:30:45.107 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[49228]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[57405]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[57405]: from='osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T11:30:45.402 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[57405]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T11:30:45.402 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:30:45 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0[61331]: 2026-03-10T11:30:45.175+0000 7f8b9cb05740 -1 osd.0 0 log_to_monitors true 2026-03-10T11:30:45.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:45.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:45.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:45.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[49228]: from='osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T11:30:45.689 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:45 vm06 ceph-mon[49228]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T11:30:45.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:45 vm09 ceph-mon[54793]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:45.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:45 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:45 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:45 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:45.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:45 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:45 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:45 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:45 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:45 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:45 vm09 ceph-mon[54793]: from='osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T11:30:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:45 vm09 ceph-mon[54793]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T11:30:46.436 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:46.452 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch daemon add osd vm06:/dev/vdd 2026-03-10T11:30:46.614 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:46.994 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: Detected new or changed devices on vm06 2026-03-10T11:30:46.994 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:46.994 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:46.994 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:46.994 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:46.994 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: from='osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[49228]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: Detected new or changed devices on vm06 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: from='osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:46.995 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:46 vm06 ceph-mon[57405]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: Detected new or changed devices on vm06 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: from='osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:47.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:46 vm09 ceph-mon[54793]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:47.270 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:30:46 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0[61331]: 2026-03-10T11:30:46.993+0000 7f8b98a86640 -1 osd.0 0 waiting for initial osdmap 2026-03-10T11:30:47.270 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:30:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0[61331]: 2026-03-10T11:30:47.002+0000 7f8b948b0640 -1 osd.0 7 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: osdmap e7: 1 total, 0 up, 1 in 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2441986695' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f6896961-eee9-4f03-8e69-efccff3a919c"}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2441986695' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f6896961-eee9-4f03-8e69-efccff3a919c"}]': finished 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574] boot 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: osdmap e8: 2 total, 1 up, 2 in 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: osdmap e7: 1 total, 0 up, 1 in 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2441986695' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f6896961-eee9-4f03-8e69-efccff3a919c"}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2441986695' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f6896961-eee9-4f03-8e69-efccff3a919c"}]': finished 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574] boot 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: osdmap e8: 2 total, 1 up, 2 in 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:48.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T11:30:48.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:48.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:30:48.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:30:48.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:48.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:30:48.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: osdmap e7: 1 total, 0 up, 1 in 2026-03-10T11:30:48.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:48.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:48.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2441986695' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f6896961-eee9-4f03-8e69-efccff3a919c"}]: dispatch 2026-03-10T11:30:48.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2441986695' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f6896961-eee9-4f03-8e69-efccff3a919c"}]': finished 2026-03-10T11:30:48.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574] boot 2026-03-10T11:30:48.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: osdmap e8: 2 total, 1 up, 2 in 2026-03-10T11:30:48.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:30:48.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:49.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:48 vm06 ceph-mon[49228]: purged_snaps scrub starts 2026-03-10T11:30:49.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:48 vm06 ceph-mon[49228]: purged_snaps scrub ok 2026-03-10T11:30:49.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:48 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/4188235588' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:30:49.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:48 vm06 ceph-mon[57405]: purged_snaps scrub starts 2026-03-10T11:30:49.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:48 vm06 ceph-mon[57405]: purged_snaps scrub ok 2026-03-10T11:30:49.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:48 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/4188235588' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:30:49.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:48 vm09 ceph-mon[54793]: purged_snaps scrub starts 2026-03-10T11:30:49.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:48 vm09 ceph-mon[54793]: purged_snaps scrub ok 2026-03-10T11:30:49.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:48 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/4188235588' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:30:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:49 vm06 ceph-mon[49228]: pgmap v18: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:49 vm06 ceph-mon[49228]: osdmap e9: 2 total, 1 up, 2 in 2026-03-10T11:30:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:49 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:49 vm06 ceph-mon[57405]: pgmap v18: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:49 vm06 ceph-mon[57405]: osdmap e9: 2 total, 1 up, 2 in 2026-03-10T11:30:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:49 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:50.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:49 vm09 ceph-mon[54793]: pgmap v18: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:50.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:49 vm09 ceph-mon[54793]: osdmap e9: 2 total, 1 up, 2 in 2026-03-10T11:30:50.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:49 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:51.852 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:51 vm06 ceph-mon[49228]: pgmap v20: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:51.852 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:51 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T11:30:51.852 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:51 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:51.852 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:51 vm06 ceph-mon[57405]: pgmap v20: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:51.852 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:51 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T11:30:51.852 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:51 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:52.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:51 vm09 ceph-mon[54793]: pgmap v20: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:52.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:51 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T11:30:52.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:51 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:52 vm06 ceph-mon[49228]: Deploying daemon osd.1 on vm06 2026-03-10T11:30:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:52 vm06 ceph-mon[57405]: Deploying daemon osd.1 on vm06 2026-03-10T11:30:53.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:52 vm09 ceph-mon[54793]: Deploying daemon osd.1 on vm06 2026-03-10T11:30:54.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:53 vm06 ceph-mon[57405]: pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:54.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:53 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:54.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:53 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:54.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:53 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:54.032 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:53 vm06 ceph-mon[49228]: pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:54.032 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:53 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:54.032 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:53 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:54.032 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:53 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:54.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:53 vm09 ceph-mon[54793]: pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:54.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:53 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:54.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:53 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:54.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:53 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:54.528 INFO:teuthology.orchestra.run.vm06.stdout:Created osd(s) 1 on host 'vm06' 2026-03-10T11:30:54.579 DEBUG:teuthology.orchestra.run.vm06:osd.1> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.1.service 2026-03-10T11:30:54.581 INFO:tasks.cephadm:Deploying osd.2 on vm06 with /dev/vdc... 2026-03-10T11:30:54.581 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- lvm zap /dev/vdc 2026-03-10T11:30:54.760 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:30:54 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1[66565]: 2026-03-10T11:30:54.695+0000 7f2f568e7740 -1 Falling back to public interface 2026-03-10T11:30:54.843 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:55.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:55 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:55 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:55 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:55.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:55 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:55.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:55 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:55 vm09 ceph-mon[54793]: pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:55.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:55 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:55.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:55 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:55 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[49228]: pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[57405]: pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.486 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:55 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:55.748 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:30:55 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1[66565]: 2026-03-10T11:30:55.558+0000 7f2f568e7740 -1 osd.1 0 log_to_monitors true 2026-03-10T11:30:56.146 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:30:56.163 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch daemon add osd vm06:/dev/vdc 2026-03-10T11:30:56.324 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:30:56.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:56 vm09 ceph-mon[54793]: Detected new or changed devices on vm06 2026-03-10T11:30:56.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:56 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:56 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:56 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:56 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:56 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:56 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:56 vm09 ceph-mon[54793]: from='osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T11:30:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:56 vm09 ceph-mon[54793]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[49228]: Detected new or changed devices on vm06 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[49228]: from='osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[49228]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[57405]: Detected new or changed devices on vm06 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[57405]: from='osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T11:30:56.739 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:56 vm06 ceph-mon[57405]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: osdmap e10: 2 total, 1 up, 2 in 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='client.24152 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2264282643' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "600ee862-9ac1-423d-83b5-9b3e3288f4f1"}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "600ee862-9ac1-423d-83b5-9b3e3288f4f1"}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "600ee862-9ac1-423d-83b5-9b3e3288f4f1"}]': finished 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: osdmap e11: 3 total, 1 up, 3 in 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: osdmap e10: 2 total, 1 up, 2 in 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='client.24152 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2264282643' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "600ee862-9ac1-423d-83b5-9b3e3288f4f1"}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "600ee862-9ac1-423d-83b5-9b3e3288f4f1"}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "600ee862-9ac1-423d-83b5-9b3e3288f4f1"}]': finished 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: osdmap e11: 3 total, 1 up, 3 in 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:30:57.694 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:57 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:57.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:57.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T11:30:57.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: osdmap e10: 2 total, 1 up, 2 in 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='client.24152 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2264282643' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "600ee862-9ac1-423d-83b5-9b3e3288f4f1"}]: dispatch 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "600ee862-9ac1-423d-83b5-9b3e3288f4f1"}]: dispatch 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "600ee862-9ac1-423d-83b5-9b3e3288f4f1"}]': finished 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: osdmap e11: 3 total, 1 up, 3 in 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:30:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:57 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:58.447 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:30:58 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1[66565]: 2026-03-10T11:30:58.173+0000 7f2f5307b640 -1 osd.1 0 waiting for initial osdmap 2026-03-10T11:30:58.447 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:30:58 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1[66565]: 2026-03-10T11:30:58.182+0000 7f2f4de91640 -1 osd.1 11 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:30:58.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:58 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2500559028' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:30:58.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:58 vm09 ceph-mon[54793]: from='osd.1 ' entity='osd.1' 2026-03-10T11:30:58.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:58 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:58.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:58 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2500559028' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:30:58.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:58 vm06 ceph-mon[49228]: from='osd.1 ' entity='osd.1' 2026-03-10T11:30:58.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:58 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:58.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:58 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2500559028' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:30:58.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:58 vm06 ceph-mon[57405]: from='osd.1 ' entity='osd.1' 2026-03-10T11:30:58.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:58 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:59.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:59 vm09 ceph-mon[54793]: purged_snaps scrub starts 2026-03-10T11:30:59.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:59 vm09 ceph-mon[54793]: purged_snaps scrub ok 2026-03-10T11:30:59.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:59 vm09 ceph-mon[54793]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:59.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:59 vm09 ceph-mon[54793]: osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054] boot 2026-03-10T11:30:59.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:59 vm09 ceph-mon[54793]: osdmap e12: 3 total, 2 up, 3 in 2026-03-10T11:30:59.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:59 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:59.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:30:59 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:30:59.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[49228]: purged_snaps scrub starts 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[49228]: purged_snaps scrub ok 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[49228]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[49228]: osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054] boot 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[49228]: osdmap e12: 3 total, 2 up, 3 in 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[57405]: purged_snaps scrub starts 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[57405]: purged_snaps scrub ok 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[57405]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[57405]: osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054] boot 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[57405]: osdmap e12: 3 total, 2 up, 3 in 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:30:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:30:59 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:01.504 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:01 vm06 ceph-mon[49228]: pgmap v28: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:01.504 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:01 vm06 ceph-mon[49228]: osdmap e13: 3 total, 2 up, 3 in 2026-03-10T11:31:01.504 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:01 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:01.504 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:01 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T11:31:01.504 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:01 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:01.504 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:01 vm06 ceph-mon[57405]: pgmap v28: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:01.504 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:01 vm06 ceph-mon[57405]: osdmap e13: 3 total, 2 up, 3 in 2026-03-10T11:31:01.504 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:01 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:01.504 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:01 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T11:31:01.504 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:01 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:01.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:01 vm09 ceph-mon[54793]: pgmap v28: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:01.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:01 vm09 ceph-mon[54793]: osdmap e13: 3 total, 2 up, 3 in 2026-03-10T11:31:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:01 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:01 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T11:31:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:01 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:02.700 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:02 vm06 ceph-mon[49228]: Deploying daemon osd.2 on vm06 2026-03-10T11:31:02.700 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:02 vm06 ceph-mon[57405]: Deploying daemon osd.2 on vm06 2026-03-10T11:31:02.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:02 vm09 ceph-mon[54793]: Deploying daemon osd.2 on vm06 2026-03-10T11:31:03.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:03 vm06 ceph-mon[49228]: pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:03.467 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:03.467 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:03 vm06 ceph-mon[57405]: pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:03.467 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:03.467 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:03.467 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:03.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:03 vm09 ceph-mon[54793]: pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:03.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:03.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:03.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:03.751 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:03.751 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:04.358 INFO:teuthology.orchestra.run.vm06.stdout:Created osd(s) 2 on host 'vm06' 2026-03-10T11:31:04.420 DEBUG:teuthology.orchestra.run.vm06:osd.2> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.2.service 2026-03-10T11:31:04.421 INFO:tasks.cephadm:Deploying osd.3 on vm06 with /dev/vdb... 2026-03-10T11:31:04.421 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- lvm zap /dev/vdb 2026-03-10T11:31:04.722 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:31:04.974 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:31:04 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2[71586]: 2026-03-10T11:31:04.880+0000 7f4b5afac740 -1 osd.2 0 log_to_monitors true 2026-03-10T11:31:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:04 vm09 ceph-mon[54793]: pgmap v31: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:04 vm09 ceph-mon[54793]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T11:31:05.265 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.265 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.265 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:05.265 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:05.265 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.265 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[49228]: pgmap v31: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:05.265 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:05.265 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.265 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.265 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[49228]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T11:31:05.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:05.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:05.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[57405]: pgmap v31: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:05.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:05.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:05.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:04 vm06 ceph-mon[57405]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T11:31:06.237 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:31:06.253 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch daemon add osd vm06:/dev/vdb 2026-03-10T11:31:06.431 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:31:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[49228]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T11:31:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[49228]: osdmap e14: 3 total, 2 up, 3 in 2026-03-10T11:31:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[49228]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:31:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[49228]: Detected new or changed devices on vm06 2026-03-10T11:31:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[57405]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[57405]: osdmap e14: 3 total, 2 up, 3 in 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[57405]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[57405]: Detected new or changed devices on vm06 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:06.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:06 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:06 vm09 ceph-mon[54793]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T11:31:07.027 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:06 vm09 ceph-mon[54793]: osdmap e14: 3 total, 2 up, 3 in 2026-03-10T11:31:07.027 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:06 vm09 ceph-mon[54793]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:31:07.027 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:06 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:07.027 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:06 vm09 ceph-mon[54793]: Detected new or changed devices on vm06 2026-03-10T11:31:07.027 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:06 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:07.028 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:06 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:07.028 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:06 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:07.028 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:06 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:07.028 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:06 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:07.028 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:06 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[57405]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[57405]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[57405]: osdmap e15: 3 total, 2 up, 3 in 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[57405]: from='client.24160 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[49228]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[49228]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[49228]: osdmap e15: 3 total, 2 up, 3 in 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[49228]: from='client.24160 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:07.954 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:07 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:07 vm09 ceph-mon[54793]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:07 vm09 ceph-mon[54793]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:31:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:07 vm09 ceph-mon[54793]: osdmap e15: 3 total, 2 up, 3 in 2026-03-10T11:31:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:07 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:07 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:07 vm09 ceph-mon[54793]: from='client.24160 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:07 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:07 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:07 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:08.233 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:31:08 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2[71586]: 2026-03-10T11:31:08.140+0000 7f4b57740640 -1 osd.2 0 waiting for initial osdmap 2026-03-10T11:31:08.233 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:31:08 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2[71586]: 2026-03-10T11:31:08.149+0000 7f4b52556640 -1 osd.2 15 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: purged_snaps scrub starts 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: purged_snaps scrub ok 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2448255248' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "80cdadbb-309f-49c9-a457-c6534574dd5a"}]: dispatch 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "80cdadbb-309f-49c9-a457-c6534574dd5a"}]: dispatch 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "80cdadbb-309f-49c9-a457-c6534574dd5a"}]': finished 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931] boot 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: osdmap e16: 4 total, 3 up, 4 in 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:08 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: purged_snaps scrub starts 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: purged_snaps scrub ok 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2448255248' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "80cdadbb-309f-49c9-a457-c6534574dd5a"}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "80cdadbb-309f-49c9-a457-c6534574dd5a"}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "80cdadbb-309f-49c9-a457-c6534574dd5a"}]': finished 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931] boot 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: osdmap e16: 4 total, 3 up, 4 in 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: purged_snaps scrub starts 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: purged_snaps scrub ok 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: from='osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931]' entity='osd.2' 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2448255248' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "80cdadbb-309f-49c9-a457-c6534574dd5a"}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "80cdadbb-309f-49c9-a457-c6534574dd5a"}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "80cdadbb-309f-49c9-a457-c6534574dd5a"}]': finished 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931] boot 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: osdmap e16: 4 total, 3 up, 4 in 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:08 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:31:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:09 vm09 ceph-mon[54793]: pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:09 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/46597361' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:09 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T11:31:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:09 vm09 ceph-mon[54793]: osdmap e17: 4 total, 3 up, 4 in 2026-03-10T11:31:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:09 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:09 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[49228]: pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/46597361' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[49228]: osdmap e17: 4 total, 3 up, 4 in 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[57405]: pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/46597361' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[57405]: osdmap e17: 4 total, 3 up, 4 in 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:09 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:31:10.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75814]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T11:31:10.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75814]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T11:31:10.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75814]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T11:31:10.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75814]: pam_unix(sudo:session): session closed for user root 2026-03-10T11:31:10.783 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75818]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T11:31:10.783 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75818]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T11:31:10.783 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75818]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T11:31:10.783 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75818]: pam_unix(sudo:session): session closed for user root 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75618]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75618]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75618]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75618]: pam_unix(sudo:session): session closed for user root 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75622]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vdd 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75622]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75622]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75622]: pam_unix(sudo:session): session closed for user root 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75630]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vdc 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75630]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75630]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T11:31:10.783 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:31:10 vm06 sudo[75630]: pam_unix(sudo:session): session closed for user root 2026-03-10T11:31:10.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:10 vm09 sudo[57151]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T11:31:10.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:10 vm09 sudo[57151]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T11:31:10.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:10 vm09 sudo[57151]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T11:31:10.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:10 vm09 sudo[57151]: pam_unix(sudo:session): session closed for user root 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: pgmap v38: 1 pgs: 1 creating+peering; 0 B data, 403 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: osdmap e18: 4 total, 3 up, 4 in 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[49228]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: pgmap v38: 1 pgs: 1 creating+peering; 0 B data, 403 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: osdmap e18: 4 total, 3 up, 4 in 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T11:31:11.622 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:11 vm06 ceph-mon[57405]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: pgmap v38: 1 pgs: 1 creating+peering; 0 B data, 403 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: osdmap e18: 4 total, 3 up, 4 in 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T11:31:11.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:11 vm09 ceph-mon[54793]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:31:12.576 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:12 vm06 ceph-mon[49228]: osdmap e19: 4 total, 3 up, 4 in 2026-03-10T11:31:12.576 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:12 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:12.830 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:12 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T11:31:12.830 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:12 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:12.830 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:12 vm06 ceph-mon[57405]: osdmap e19: 4 total, 3 up, 4 in 2026-03-10T11:31:12.830 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:12 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:12.830 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:12 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T11:31:12.830 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:12 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:12.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:12 vm09 ceph-mon[54793]: osdmap e19: 4 total, 3 up, 4 in 2026-03-10T11:31:12.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:12 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:12.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:12 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T11:31:12.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:12 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:13 vm06 ceph-mon[49228]: pgmap v41: 1 pgs: 1 creating+peering; 0 B data, 403 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:13 vm06 ceph-mon[49228]: Deploying daemon osd.3 on vm06 2026-03-10T11:31:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:13 vm06 ceph-mon[49228]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-10T11:31:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:13 vm06 ceph-mon[49228]: Cluster is now healthy 2026-03-10T11:31:13.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:13 vm06 ceph-mon[49228]: mgrmap e15: y(active, since 64s), standbys: x 2026-03-10T11:31:13.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:13 vm06 ceph-mon[57405]: pgmap v41: 1 pgs: 1 creating+peering; 0 B data, 403 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:13.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:13 vm06 ceph-mon[57405]: Deploying daemon osd.3 on vm06 2026-03-10T11:31:13.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:13 vm06 ceph-mon[57405]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-10T11:31:13.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:13 vm06 ceph-mon[57405]: Cluster is now healthy 2026-03-10T11:31:13.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:13 vm06 ceph-mon[57405]: mgrmap e15: y(active, since 64s), standbys: x 2026-03-10T11:31:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:13 vm09 ceph-mon[54793]: pgmap v41: 1 pgs: 1 creating+peering; 0 B data, 403 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:13 vm09 ceph-mon[54793]: Deploying daemon osd.3 on vm06 2026-03-10T11:31:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:13 vm09 ceph-mon[54793]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-10T11:31:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:13 vm09 ceph-mon[54793]: Cluster is now healthy 2026-03-10T11:31:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:13 vm09 ceph-mon[54793]: mgrmap e15: y(active, since 64s), standbys: x 2026-03-10T11:31:14.834 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:14.834 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:14.834 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:14 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:14.834 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:14 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:14.834 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:14 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:14.834 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:14 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:14.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:14 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:14.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:14 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:14.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:14 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:15.726 INFO:teuthology.orchestra.run.vm06.stdout:Created osd(s) 3 on host 'vm06' 2026-03-10T11:31:15.783 DEBUG:teuthology.orchestra.run.vm06:osd.3> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.3.service 2026-03-10T11:31:15.787 INFO:tasks.cephadm:Deploying osd.4 on vm09 with /dev/vde... 2026-03-10T11:31:15.787 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- lvm zap /dev/vde 2026-03-10T11:31:15.844 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:15 vm06 ceph-mon[49228]: pgmap v42: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:15.844 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:15 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:15.844 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:15 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:15.845 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:15 vm06 ceph-mon[57405]: pgmap v42: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:15.845 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:15 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:15.845 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:15 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:15.953 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:31:15.977 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:15 vm09 ceph-mon[54793]: pgmap v42: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:15.977 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:15 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:15.977 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:15 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:16.532 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:31:16 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3[76780]: 2026-03-10T11:31:16.141+0000 7ff889ecf740 -1 osd.3 0 log_to_monitors true 2026-03-10T11:31:16.770 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:31:16.788 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch daemon add osd vm09:/dev/vde 2026-03-10T11:31:16.963 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:31:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:16 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:16 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:16 vm09 ceph-mon[54793]: from='osd.3 [v2:192.168.123.106:6826/1874161814,v1:192.168.123.106:6827/1874161814]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T11:31:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:16 vm09 ceph-mon[54793]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T11:31:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:16 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:16 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:16 vm06 ceph-mon[49228]: from='osd.3 [v2:192.168.123.106:6826/1874161814,v1:192.168.123.106:6827/1874161814]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T11:31:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:16 vm06 ceph-mon[49228]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T11:31:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:16 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:16 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:16 vm06 ceph-mon[57405]: from='osd.3 [v2:192.168.123.106:6826/1874161814,v1:192.168.123.106:6827/1874161814]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T11:31:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:16 vm06 ceph-mon[57405]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: pgmap v43: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: osdmap e20: 4 total, 3 up, 4 in 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='osd.3 [v2:192.168.123.106:6826/1874161814,v1:192.168.123.106:6827/1874161814]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: Detected new or changed devices on vm06 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: pgmap v43: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: osdmap e20: 4 total, 3 up, 4 in 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='osd.3 [v2:192.168.123.106:6826/1874161814,v1:192.168.123.106:6827/1874161814]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: Detected new or changed devices on vm06 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: pgmap v43: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: osdmap e20: 4 total, 3 up, 4 in 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='osd.3 [v2:192.168.123.106:6826/1874161814,v1:192.168.123.106:6827/1874161814]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: Detected new or changed devices on vm06 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:18.032 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:18.032 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:18.032 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:18.032 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:18.032 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:18.032 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:18.032 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:18.032 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:18.032 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:31:17 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3[76780]: 2026-03-10T11:31:17.739+0000 7ff885e50640 -1 osd.3 0 waiting for initial osdmap 2026-03-10T11:31:18.032 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:31:17 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3[76780]: 2026-03-10T11:31:17.745+0000 7ff881c7a640 -1 osd.3 21 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: from='client.14298 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: osdmap e21: 4 total, 3 up, 4 in 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/2128172286' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3f0439f8-e455-4db3-88a6-7812e1b40aa7"}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3f0439f8-e455-4db3-88a6-7812e1b40aa7"}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "3f0439f8-e455-4db3-88a6-7812e1b40aa7"}]': finished 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: osd.3 [v2:192.168.123.106:6826/1874161814,v1:192.168.123.106:6827/1874161814] boot 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: osdmap e22: 5 total, 4 up, 5 in 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/2868735083' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: from='client.14298 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: osdmap e21: 4 total, 3 up, 4 in 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/2128172286' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3f0439f8-e455-4db3-88a6-7812e1b40aa7"}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3f0439f8-e455-4db3-88a6-7812e1b40aa7"}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "3f0439f8-e455-4db3-88a6-7812e1b40aa7"}]': finished 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: osd.3 [v2:192.168.123.106:6826/1874161814,v1:192.168.123.106:6827/1874161814] boot 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: osdmap e22: 5 total, 4 up, 5 in 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:18 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/2868735083' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: from='client.14298 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: osdmap e21: 4 total, 3 up, 4 in 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/2128172286' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3f0439f8-e455-4db3-88a6-7812e1b40aa7"}]: dispatch 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3f0439f8-e455-4db3-88a6-7812e1b40aa7"}]: dispatch 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "3f0439f8-e455-4db3-88a6-7812e1b40aa7"}]': finished 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: osd.3 [v2:192.168.123.106:6826/1874161814,v1:192.168.123.106:6827/1874161814] boot 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: osdmap e22: 5 total, 4 up, 5 in 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:18 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/2868735083' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:20.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:19 vm09 ceph-mon[54793]: purged_snaps scrub starts 2026-03-10T11:31:20.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:19 vm09 ceph-mon[54793]: purged_snaps scrub ok 2026-03-10T11:31:20.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:19 vm09 ceph-mon[54793]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:20.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:19 vm09 ceph-mon[54793]: osdmap e23: 5 total, 4 up, 5 in 2026-03-10T11:31:20.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:19 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:20.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:19 vm06 ceph-mon[49228]: purged_snaps scrub starts 2026-03-10T11:31:20.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:19 vm06 ceph-mon[49228]: purged_snaps scrub ok 2026-03-10T11:31:20.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:19 vm06 ceph-mon[49228]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:20.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:19 vm06 ceph-mon[49228]: osdmap e23: 5 total, 4 up, 5 in 2026-03-10T11:31:20.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:19 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:20.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:19 vm06 ceph-mon[57405]: purged_snaps scrub starts 2026-03-10T11:31:20.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:19 vm06 ceph-mon[57405]: purged_snaps scrub ok 2026-03-10T11:31:20.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:19 vm06 ceph-mon[57405]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T11:31:20.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:19 vm06 ceph-mon[57405]: osdmap e23: 5 total, 4 up, 5 in 2026-03-10T11:31:20.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:19 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:22.117 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:21 vm09 ceph-mon[54793]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:22.117 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:21 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T11:31:22.117 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:21 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:22.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:21 vm06 ceph-mon[49228]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:22.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:21 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T11:31:22.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:21 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:22.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:21 vm06 ceph-mon[57405]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:22.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:21 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T11:31:22.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:21 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:23.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:22 vm09 ceph-mon[54793]: Deploying daemon osd.4 on vm09 2026-03-10T11:31:23.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:22 vm06 ceph-mon[49228]: Deploying daemon osd.4 on vm09 2026-03-10T11:31:23.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:22 vm06 ceph-mon[57405]: Deploying daemon osd.4 on vm09 2026-03-10T11:31:24.051 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:24 vm09 ceph-mon[54793]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:24.051 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:24.051 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:24.051 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:24 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:24 vm06 ceph-mon[49228]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:24 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:24 vm06 ceph-mon[57405]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:24 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:24 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:24 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:24.656 INFO:teuthology.orchestra.run.vm09.stdout:Created osd(s) 4 on host 'vm09' 2026-03-10T11:31:24.709 DEBUG:teuthology.orchestra.run.vm09:osd.4> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.4.service 2026-03-10T11:31:24.711 INFO:tasks.cephadm:Deploying osd.5 on vm09 with /dev/vdd... 2026-03-10T11:31:24.711 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- lvm zap /dev/vdd 2026-03-10T11:31:24.988 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:31:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:25 vm09 ceph-mon[54793]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:25 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:25 vm09 ceph-mon[54793]: from='osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T11:31:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:25 vm09 ceph-mon[54793]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T11:31:25.480 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:31:25 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4[59293]: 2026-03-10T11:31:25.218+0000 7f5312d65740 -1 osd.4 0 log_to_monitors true 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[49228]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[49228]: from='osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[49228]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[57405]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[57405]: from='osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T11:31:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:25 vm06 ceph-mon[57405]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T11:31:26.270 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:31:26.285 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch daemon add osd vm09:/dev/vdd 2026-03-10T11:31:26.441 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:31:26.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: Detected new or changed devices on vm09 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: Adjusting osd_memory_target on vm09 to 257.0M 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: Unable to set osd_memory_target on vm09 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: osdmap e24: 5 total, 4 up, 5 in 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: from='osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:26.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:26 vm09 ceph-mon[54793]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: Detected new or changed devices on vm09 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: Adjusting osd_memory_target on vm09 to 257.0M 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: Unable to set osd_memory_target on vm09 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: osdmap e24: 5 total, 4 up, 5 in 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: from='osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[49228]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: Detected new or changed devices on vm09 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: Adjusting osd_memory_target on vm09 to 257.0M 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: Unable to set osd_memory_target on vm09 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: osdmap e24: 5 total, 4 up, 5 in 2026-03-10T11:31:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:27.032 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: from='osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:27.032 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:26 vm06 ceph-mon[57405]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:27.917 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:31:27 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4[59293]: 2026-03-10T11:31:27.705+0000 7f530f4f9640 -1 osd.4 0 waiting for initial osdmap 2026-03-10T11:31:27.917 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:31:27 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4[59293]: 2026-03-10T11:31:27.719+0000 7f530a30f640 -1 osd.4 26 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:31:27.917 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:27.917 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:31:27.917 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: osdmap e25: 5 total, 4 up, 5 in 2026-03-10T11:31:27.917 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='client.24220 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:27.917 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:27.917 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:27.917 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:27.917 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:27.918 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:27.918 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/3177883343' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "57003e78-f740-4009-98e4-a399b0d4d62a"}]: dispatch 2026-03-10T11:31:27.918 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "57003e78-f740-4009-98e4-a399b0d4d62a"}]: dispatch 2026-03-10T11:31:27.918 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "57003e78-f740-4009-98e4-a399b0d4d62a"}]': finished 2026-03-10T11:31:27.918 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: osdmap e26: 6 total, 4 up, 6 in 2026-03-10T11:31:27.918 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:27.918 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:27 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: osdmap e25: 5 total, 4 up, 5 in 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='client.24220 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/3177883343' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "57003e78-f740-4009-98e4-a399b0d4d62a"}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "57003e78-f740-4009-98e4-a399b0d4d62a"}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "57003e78-f740-4009-98e4-a399b0d4d62a"}]': finished 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: osdmap e26: 6 total, 4 up, 6 in 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: osdmap e25: 5 total, 4 up, 5 in 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='client.24220 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/3177883343' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "57003e78-f740-4009-98e4-a399b0d4d62a"}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "57003e78-f740-4009-98e4-a399b0d4d62a"}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "57003e78-f740-4009-98e4-a399b0d4d62a"}]': finished 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: osdmap e26: 6 total, 4 up, 6 in 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:27 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:28.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:28 vm09 ceph-mon[54793]: purged_snaps scrub starts 2026-03-10T11:31:28.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:28 vm09 ceph-mon[54793]: purged_snaps scrub ok 2026-03-10T11:31:28.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:28 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:28.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:28 vm09 ceph-mon[54793]: from='osd.4 ' entity='osd.4' 2026-03-10T11:31:28.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:28 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/2321219758' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:28.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:28 vm09 ceph-mon[54793]: osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642] boot 2026-03-10T11:31:28.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:28 vm09 ceph-mon[54793]: osdmap e27: 6 total, 5 up, 6 in 2026-03-10T11:31:28.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:28 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:28.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:28 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[49228]: purged_snaps scrub starts 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[49228]: purged_snaps scrub ok 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[49228]: from='osd.4 ' entity='osd.4' 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/2321219758' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[49228]: osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642] boot 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[49228]: osdmap e27: 6 total, 5 up, 6 in 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[57405]: purged_snaps scrub starts 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[57405]: purged_snaps scrub ok 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[57405]: from='osd.4 ' entity='osd.4' 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/2321219758' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[57405]: osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642] boot 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[57405]: osdmap e27: 6 total, 5 up, 6 in 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:31:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:28 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:29.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:29 vm09 ceph-mon[54793]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:29.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:29 vm09 ceph-mon[54793]: osdmap e28: 6 total, 5 up, 6 in 2026-03-10T11:31:29.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:29 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:30.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:29 vm06 ceph-mon[49228]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:30.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:29 vm06 ceph-mon[49228]: osdmap e28: 6 total, 5 up, 6 in 2026-03-10T11:31:30.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:29 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:30.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:29 vm06 ceph-mon[57405]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T11:31:30.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:29 vm06 ceph-mon[57405]: osdmap e28: 6 total, 5 up, 6 in 2026-03-10T11:31:30.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:29 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:31.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:31 vm09 ceph-mon[54793]: pgmap v59: 1 pgs: 1 remapped+peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-10T11:31:31.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:31 vm09 ceph-mon[54793]: osdmap e29: 6 total, 5 up, 6 in 2026-03-10T11:31:31.707 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:31 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:31 vm06 ceph-mon[49228]: pgmap v59: 1 pgs: 1 remapped+peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-10T11:31:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:31 vm06 ceph-mon[49228]: osdmap e29: 6 total, 5 up, 6 in 2026-03-10T11:31:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:31 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:31.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:31 vm06 ceph-mon[57405]: pgmap v59: 1 pgs: 1 remapped+peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-10T11:31:31.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:31 vm06 ceph-mon[57405]: osdmap e29: 6 total, 5 up, 6 in 2026-03-10T11:31:31.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:31 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:32.569 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:32 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T11:31:32.570 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:32 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:32.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:32 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T11:31:32.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:32 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:32 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T11:31:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:32 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:33.764 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:33 vm09 ceph-mon[54793]: Deploying daemon osd.5 on vm09 2026-03-10T11:31:33.764 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:33 vm09 ceph-mon[54793]: pgmap v61: 1 pgs: 1 remapped+peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-10T11:31:33.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:33 vm06 ceph-mon[49228]: Deploying daemon osd.5 on vm09 2026-03-10T11:31:33.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:33 vm06 ceph-mon[49228]: pgmap v61: 1 pgs: 1 remapped+peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-10T11:31:33.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:33 vm06 ceph-mon[57405]: Deploying daemon osd.5 on vm09 2026-03-10T11:31:33.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:33 vm06 ceph-mon[57405]: pgmap v61: 1 pgs: 1 remapped+peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-10T11:31:34.756 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:34.757 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:34.757 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:34 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:34.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:34.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:34.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:34 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:34.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:34.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:34.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:34 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:35.426 INFO:teuthology.orchestra.run.vm09.stdout:Created osd(s) 5 on host 'vm09' 2026-03-10T11:31:35.486 DEBUG:teuthology.orchestra.run.vm09:osd.5> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.5.service 2026-03-10T11:31:35.488 INFO:tasks.cephadm:Deploying osd.6 on vm09 with /dev/vdc... 2026-03-10T11:31:35.488 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- lvm zap /dev/vdc 2026-03-10T11:31:35.733 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:31:36.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:35 vm09 ceph-mon[54793]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T11:31:36.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:36.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:36.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:36.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:35 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.232 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:31:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:31:35.953+0000 7f578e109740 -1 osd.5 0 log_to_monitors true 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[49228]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[57405]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:35 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:37.209 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:36 vm09 ceph-mon[54793]: from='osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T11:31:37.209 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:36 vm09 ceph-mon[54793]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T11:31:37.209 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:37.209 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:37.209 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:37.209 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:37.209 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:37.209 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:37.209 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:36 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:37.247 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:31:37.271 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch daemon add osd vm09:/dev/vdc 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[49228]: from='osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[49228]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[57405]: from='osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[57405]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:36 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:37.461 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 58 KiB/s, 0 objects/s recovering 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: Detected new or changed devices on vm09 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: Adjusting osd_memory_target on vm09 to 128.5M 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: Unable to set osd_memory_target on vm09 to 134768230: error parsing value: Value '134768230' is below minimum 939524096 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: from='osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: osdmap e30: 6 total, 5 up, 6 in 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:37 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:38.230 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:31:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:31:37.972+0000 7f578a89d640 -1 osd.5 0 waiting for initial osdmap 2026-03-10T11:31:38.230 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:31:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:31:37.984+0000 7f5785eb4640 -1 osd.5 31 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 58 KiB/s, 0 objects/s recovering 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: Detected new or changed devices on vm09 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: Adjusting osd_memory_target on vm09 to 128.5M 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: Unable to set osd_memory_target on vm09 to 134768230: error parsing value: Value '134768230' is below minimum 939524096 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: from='osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: osdmap e30: 6 total, 5 up, 6 in 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 58 KiB/s, 0 objects/s recovering 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: Detected new or changed devices on vm09 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: Adjusting osd_memory_target on vm09 to 128.5M 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: Unable to set osd_memory_target on vm09 to 134768230: error parsing value: Value '134768230' is below minimum 939524096 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: from='osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: osdmap e30: 6 total, 5 up, 6 in 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:38.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:37 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: purged_snaps scrub starts 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: purged_snaps scrub ok 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: from='client.24247 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: osdmap e31: 6 total, 5 up, 6 in 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: pgmap v66: 1 pgs: 1 unknown; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179] boot 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: osdmap e32: 6 total, 6 up, 6 in 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/2098157864' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111"}]: dispatch 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111"}]: dispatch 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111"}]': finished 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: osdmap e33: 7 total, 6 up, 7 in 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: purged_snaps scrub starts 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: purged_snaps scrub ok 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: from='client.24247 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: osdmap e31: 6 total, 5 up, 6 in 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: pgmap v66: 1 pgs: 1 unknown; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179] boot 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: osdmap e32: 6 total, 6 up, 6 in 2026-03-10T11:31:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:39.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/2098157864' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111"}]: dispatch 2026-03-10T11:31:39.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111"}]: dispatch 2026-03-10T11:31:39.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111"}]': finished 2026-03-10T11:31:39.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: osdmap e33: 7 total, 6 up, 7 in 2026-03-10T11:31:39.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:38 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: purged_snaps scrub starts 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: purged_snaps scrub ok 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: from='client.24247 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: osdmap e31: 6 total, 5 up, 6 in 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: pgmap v66: 1 pgs: 1 unknown; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179] boot 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: osdmap e32: 6 total, 6 up, 6 in 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/2098157864' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111"}]: dispatch 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111"}]: dispatch 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111"}]': finished 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: osdmap e33: 7 total, 6 up, 7 in 2026-03-10T11:31:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:38 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:40.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:39 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/1623607016' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:40.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:39 vm06 ceph-mon[49228]: osdmap e34: 7 total, 6 up, 7 in 2026-03-10T11:31:40.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:39 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:40.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:39 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/1623607016' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:40.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:39 vm06 ceph-mon[57405]: osdmap e34: 7 total, 6 up, 7 in 2026-03-10T11:31:40.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:39 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:40.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:39 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/1623607016' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:40.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:39 vm09 ceph-mon[54793]: osdmap e34: 7 total, 6 up, 7 in 2026-03-10T11:31:40.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:39 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:41.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:40 vm06 ceph-mon[49228]: pgmap v70: 1 pgs: 1 peering; 0 B data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:41.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:40 vm06 ceph-mon[57405]: pgmap v70: 1 pgs: 1 peering; 0 B data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:41.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:40 vm09 ceph-mon[54793]: pgmap v70: 1 pgs: 1 peering; 0 B data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:43.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:43 vm09 ceph-mon[54793]: pgmap v71: 1 pgs: 1 peering; 0 B data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:43.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:43 vm06 ceph-mon[49228]: pgmap v71: 1 pgs: 1 peering; 0 B data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:43.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:43 vm06 ceph-mon[57405]: pgmap v71: 1 pgs: 1 peering; 0 B data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:44.440 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:44 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T11:31:44.440 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:44 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:44.440 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:44 vm09 ceph-mon[54793]: Deploying daemon osd.6 on vm09 2026-03-10T11:31:44.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:44 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T11:31:44.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:44 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:44.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:44 vm06 ceph-mon[49228]: Deploying daemon osd.6 on vm09 2026-03-10T11:31:44.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:44 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T11:31:44.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:44 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:44.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:44 vm06 ceph-mon[57405]: Deploying daemon osd.6 on vm09 2026-03-10T11:31:45.618 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:45 vm09 ceph-mon[54793]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:45.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:45 vm06 ceph-mon[57405]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:45.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:45 vm06 ceph-mon[49228]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:46.439 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:46 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:46.439 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:46 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:46.439 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:46 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:46.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:46 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:46.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:46 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:46.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:46 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:46.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:46 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:46.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:46 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:46.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:46 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:46.975 INFO:teuthology.orchestra.run.vm09.stdout:Created osd(s) 6 on host 'vm09' 2026-03-10T11:31:47.048 DEBUG:teuthology.orchestra.run.vm09:osd.6> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.6.service 2026-03-10T11:31:47.049 INFO:tasks.cephadm:Deploying osd.7 on vm09 with /dev/vdb... 2026-03-10T11:31:47.049 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- lvm zap /dev/vdb 2026-03-10T11:31:47.359 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:31:47.523 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:31:47 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:31:47.458+0000 7fcb1c657740 -1 osd.6 0 log_to_monitors true 2026-03-10T11:31:47.926 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:47 vm09 ceph-mon[54793]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:47.926 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:47.926 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:47.926 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:47.926 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:47.926 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:47.926 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:47.926 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:47.926 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:47 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:47.926 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:47 vm09 ceph-mon[54793]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T11:31:48.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[49228]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[49228]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[57405]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:47 vm06 ceph-mon[57405]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T11:31:49.000 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:31:49.020 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch daemon add osd vm09:/dev/vdb 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: osdmap e35: 7 total, 6 up, 7 in 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: Detected new or changed devices on vm09 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: Adjusting osd_memory_target on vm09 to 87739k 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: Unable to set osd_memory_target on vm09 to 89845486: error parsing value: Value '89845486' is below minimum 939524096 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:49.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:49 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:49.216 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: osdmap e35: 7 total, 6 up, 7 in 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: Detected new or changed devices on vm09 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: Adjusting osd_memory_target on vm09 to 87739k 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: Unable to set osd_memory_target on vm09 to 89845486: error parsing value: Value '89845486' is below minimum 939524096 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: osdmap e35: 7 total, 6 up, 7 in 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: Detected new or changed devices on vm09 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:31:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: Adjusting osd_memory_target on vm09 to 87739k 2026-03-10T11:31:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: Unable to set osd_memory_target on vm09 to 89845486: error parsing value: Value '89845486' is below minimum 939524096 2026-03-10T11:31:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:49 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:50.481 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:50 vm09 ceph-mon[54793]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:31:50.481 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:50 vm09 ceph-mon[54793]: osdmap e36: 7 total, 6 up, 7 in 2026-03-10T11:31:50.481 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:50 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:50.481 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:50 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:50.481 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:50 vm09 ceph-mon[54793]: from='client.14355 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:50.481 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:50 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:50.481 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:50 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:50.481 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:50 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:50.481 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:50 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[49228]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[49228]: osdmap e36: 7 total, 6 up, 7 in 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[49228]: from='client.14355 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[57405]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[57405]: osdmap e36: 7 total, 6 up, 7 in 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[57405]: from='client.14355 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:50 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:51.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:31:51 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:31:51.023+0000 7fcb18deb640 -1 osd.6 0 waiting for initial osdmap 2026-03-10T11:31:51.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:31:51 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:31:51.041+0000 7fcb13c01640 -1 osd.6 38 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: purged_snaps scrub starts 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: purged_snaps scrub ok 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: osdmap e37: 7 total, 6 up, 7 in 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/315852219' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8b7c23f9-10d4-4419-a925-798f85675eef"}]: dispatch 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8b7c23f9-10d4-4419-a925-798f85675eef"}]: dispatch 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8b7c23f9-10d4-4419-a925-798f85675eef"}]': finished 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: osdmap e38: 8 total, 6 up, 8 in 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: purged_snaps scrub starts 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: purged_snaps scrub ok 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: osdmap e37: 7 total, 6 up, 7 in 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/315852219' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8b7c23f9-10d4-4419-a925-798f85675eef"}]: dispatch 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8b7c23f9-10d4-4419-a925-798f85675eef"}]: dispatch 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8b7c23f9-10d4-4419-a925-798f85675eef"}]': finished 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: osdmap e38: 8 total, 6 up, 8 in 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' 2026-03-10T11:31:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:51 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: purged_snaps scrub starts 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: purged_snaps scrub ok 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: osdmap e37: 7 total, 6 up, 7 in 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/315852219' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8b7c23f9-10d4-4419-a925-798f85675eef"}]: dispatch 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8b7c23f9-10d4-4419-a925-798f85675eef"}]: dispatch 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8b7c23f9-10d4-4419-a925-798f85675eef"}]': finished 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: osdmap e38: 8 total, 6 up, 8 in 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: from='osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529]' entity='osd.6' 2026-03-10T11:31:51.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:51 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:52.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:52 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/3244999891' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:52.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:52 vm06 ceph-mon[49228]: osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529] boot 2026-03-10T11:31:52.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:52 vm06 ceph-mon[49228]: osdmap e39: 8 total, 7 up, 8 in 2026-03-10T11:31:52.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:52 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:52.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:52 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:52.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:52 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/3244999891' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:52.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:52 vm06 ceph-mon[57405]: osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529] boot 2026-03-10T11:31:52.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:52 vm06 ceph-mon[57405]: osdmap e39: 8 total, 7 up, 8 in 2026-03-10T11:31:52.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:52 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:52.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:52 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:52.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:52 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/3244999891' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T11:31:52.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:52 vm09 ceph-mon[54793]: osd.6 [v2:192.168.123.109:6816/536869529,v1:192.168.123.109:6817/536869529] boot 2026-03-10T11:31:52.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:52 vm09 ceph-mon[54793]: osdmap e39: 8 total, 7 up, 8 in 2026-03-10T11:31:52.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:52 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:31:52.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:52 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:53 vm06 ceph-mon[49228]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 587 MiB used, 139 GiB / 140 GiB avail 2026-03-10T11:31:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:53 vm06 ceph-mon[49228]: osdmap e40: 8 total, 7 up, 8 in 2026-03-10T11:31:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:53 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:53 vm06 ceph-mon[57405]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 587 MiB used, 139 GiB / 140 GiB avail 2026-03-10T11:31:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:53 vm06 ceph-mon[57405]: osdmap e40: 8 total, 7 up, 8 in 2026-03-10T11:31:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:53 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:53.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:53 vm09 ceph-mon[54793]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 587 MiB used, 139 GiB / 140 GiB avail 2026-03-10T11:31:53.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:53 vm09 ceph-mon[54793]: osdmap e40: 8 total, 7 up, 8 in 2026-03-10T11:31:53.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:53 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:55.156 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:55 vm09 ceph-mon[54793]: osdmap e41: 8 total, 7 up, 8 in 2026-03-10T11:31:55.156 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:55 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:55.156 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:55 vm09 ceph-mon[54793]: pgmap v84: 1 pgs: 1 remapped+peering; 449 KiB data, 587 MiB used, 139 GiB / 140 GiB avail 2026-03-10T11:31:55.156 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:55 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T11:31:55.156 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:55 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:55.156 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:55 vm09 ceph-mon[54793]: Deploying daemon osd.7 on vm09 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[49228]: osdmap e41: 8 total, 7 up, 8 in 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[49228]: pgmap v84: 1 pgs: 1 remapped+peering; 449 KiB data, 587 MiB used, 139 GiB / 140 GiB avail 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[49228]: Deploying daemon osd.7 on vm09 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[57405]: osdmap e41: 8 total, 7 up, 8 in 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[57405]: pgmap v84: 1 pgs: 1 remapped+peering; 449 KiB data, 587 MiB used, 139 GiB / 140 GiB avail 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:55.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:55 vm06 ceph-mon[57405]: Deploying daemon osd.7 on vm09 2026-03-10T11:31:57.589 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:57 vm09 ceph-mon[54793]: pgmap v85: 1 pgs: 1 remapped+peering; 449 KiB data, 587 MiB used, 139 GiB / 140 GiB avail 2026-03-10T11:31:57.589 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:57 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:57.589 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:57 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:57.589 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:57 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:57.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:57 vm06 ceph-mon[49228]: pgmap v85: 1 pgs: 1 remapped+peering; 449 KiB data, 587 MiB used, 139 GiB / 140 GiB avail 2026-03-10T11:31:57.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:57 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:57.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:57 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:57.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:57 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:57.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:57 vm06 ceph-mon[57405]: pgmap v85: 1 pgs: 1 remapped+peering; 449 KiB data, 587 MiB used, 139 GiB / 140 GiB avail 2026-03-10T11:31:57.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:57 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:57.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:57 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:57.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:57 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:58.407 INFO:teuthology.orchestra.run.vm09.stdout:Created osd(s) 7 on host 'vm09' 2026-03-10T11:31:58.464 DEBUG:teuthology.orchestra.run.vm09:osd.7> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.7.service 2026-03-10T11:31:58.466 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-10T11:31:58.466 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd stat -f json 2026-03-10T11:31:58.666 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:31:58.905 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:31:58.961 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":41,"num_osds":8,"num_up_osds":7,"osd_up_since":1773142312,"num_in_osds":8,"osd_in_since":1773142310,"num_remapped_pgs":0} 2026-03-10T11:31:58.980 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:31:58 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:31:58.885+0000 7f509b68e740 -1 osd.7 0 log_to_monitors true 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 188 MiB used, 140 GiB / 140 GiB avail 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: from='osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T11:31:59.307 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:31:59 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2084562187' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 188 MiB used, 140 GiB / 140 GiB avail 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: from='osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2084562187' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 188 MiB used, 140 GiB / 140 GiB avail 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: from='osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T11:31:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:31:59 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2084562187' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T11:31:59.962 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd stat -f json 2026-03-10T11:32:00.143 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:00.368 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:32:00.471 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":42,"num_osds":8,"num_up_osds":7,"osd_up_since":1773142312,"num_in_osds":8,"osd_in_since":1773142310,"num_remapped_pgs":0} 2026-03-10T11:32:00.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: Detected new or changed devices on vm09 2026-03-10T11:32:00.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:00.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:00.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: Adjusting osd_memory_target on vm09 to 65804k 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: Unable to set osd_memory_target on vm09 to 67384115: error parsing value: Value '67384115' is below minimum 939524096 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: osdmap e42: 8 total, 7 up, 8 in 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2626303885' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: Detected new or changed devices on vm09 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: Adjusting osd_memory_target on vm09 to 65804k 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: Unable to set osd_memory_target on vm09 to 67384115: error parsing value: Value '67384115' is below minimum 939524096 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: osdmap e42: 8 total, 7 up, 8 in 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:32:00.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:00 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2626303885' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: Detected new or changed devices on vm09 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: Adjusting osd_memory_target on vm09 to 65804k 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: Unable to set osd_memory_target on vm09 to 67384115: error parsing value: Value '67384115' is below minimum 939524096 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: osdmap e42: 8 total, 7 up, 8 in 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T11:32:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:00 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2626303885' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T11:32:01.472 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd stat -f json 2026-03-10T11:32:01.648 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:01.674 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[49228]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 62 KiB/s, 0 objects/s recovering 2026-03-10T11:32:01.674 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[49228]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:32:01.674 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[49228]: osdmap e43: 8 total, 7 up, 8 in 2026-03-10T11:32:01.674 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:01.674 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:01.675 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[49228]: from='osd.7 ' entity='osd.7' 2026-03-10T11:32:01.675 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[57405]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 62 KiB/s, 0 objects/s recovering 2026-03-10T11:32:01.675 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[57405]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:32:01.675 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[57405]: osdmap e43: 8 total, 7 up, 8 in 2026-03-10T11:32:01.675 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:01.675 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:01.675 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:01 vm06 ceph-mon[57405]: from='osd.7 ' entity='osd.7' 2026-03-10T11:32:01.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:01 vm09 ceph-mon[54793]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail; 62 KiB/s, 0 objects/s recovering 2026-03-10T11:32:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:01 vm09 ceph-mon[54793]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T11:32:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:01 vm09 ceph-mon[54793]: osdmap e43: 8 total, 7 up, 8 in 2026-03-10T11:32:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:01 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:01 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:01 vm09 ceph-mon[54793]: from='osd.7 ' entity='osd.7' 2026-03-10T11:32:01.730 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:32:01 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:32:01.380+0000 7f509760f640 -1 osd.7 0 waiting for initial osdmap 2026-03-10T11:32:01.730 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:32:01 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:32:01.392+0000 7f5092c38640 -1 osd.7 43 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:32:01.885 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:32:01.964 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":44,"num_osds":8,"num_up_osds":8,"osd_up_since":1773142321,"num_in_osds":8,"osd_in_since":1773142310,"num_remapped_pgs":1} 2026-03-10T11:32:01.964 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd dump --format=json 2026-03-10T11:32:02.136 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:02.368 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:32:02.369 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":44,"fsid":"52119272-1c74-11f1-990e-e7cc1c0e3258","created":"2026-03-10T11:29:47.393973+0000","modified":"2026-03-10T11:32:01.407381+0000","last_up_change":"2026-03-10T11:32:01.407381+0000","last_in_change":"2026-03-10T11:31:50.876994+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T11:31:08.303126+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"19","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"7704abb8-f49c-4aef-89e1-d9e67d3e765b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":43,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6803","nonce":3972625574}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6805","nonce":3972625574}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6809","nonce":3972625574}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6807","nonce":3972625574}]},"public_addr":"192.168.123.106:6803/3972625574","cluster_addr":"192.168.123.106:6805/3972625574","heartbeat_back_addr":"192.168.123.106:6809/3972625574","heartbeat_front_addr":"192.168.123.106:6807/3972625574","state":["exists","up"]},{"osd":1,"uuid":"f6896961-eee9-4f03-8e69-efccff3a919c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":28,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6811","nonce":193031054}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6813","nonce":193031054}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6817","nonce":193031054}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6815","nonce":193031054}]},"public_addr":"192.168.123.106:6811/193031054","cluster_addr":"192.168.123.106:6813/193031054","heartbeat_back_addr":"192.168.123.106:6817/193031054","heartbeat_front_addr":"192.168.123.106:6815/193031054","state":["exists","up"]},{"osd":2,"uuid":"600ee862-9ac1-423d-83b5-9b3e3288f4f1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6819","nonce":316237931}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6821","nonce":316237931}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6825","nonce":316237931}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6823","nonce":316237931}]},"public_addr":"192.168.123.106:6819/316237931","cluster_addr":"192.168.123.106:6821/316237931","heartbeat_back_addr":"192.168.123.106:6825/316237931","heartbeat_front_addr":"192.168.123.106:6823/316237931","state":["exists","up"]},{"osd":3,"uuid":"80cdadbb-309f-49c9-a457-c6534574dd5a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":22,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6827","nonce":1874161814}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6829","nonce":1874161814}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6832","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6833","nonce":1874161814}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6831","nonce":1874161814}]},"public_addr":"192.168.123.106:6827/1874161814","cluster_addr":"192.168.123.106:6829/1874161814","heartbeat_back_addr":"192.168.123.106:6833/1874161814","heartbeat_front_addr":"192.168.123.106:6831/1874161814","state":["exists","up"]},{"osd":4,"uuid":"3f0439f8-e455-4db3-88a6-7812e1b40aa7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6801","nonce":3122480642}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6803","nonce":3122480642}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6807","nonce":3122480642}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6805","nonce":3122480642}]},"public_addr":"192.168.123.109:6801/3122480642","cluster_addr":"192.168.123.109:6803/3122480642","heartbeat_back_addr":"192.168.123.109:6807/3122480642","heartbeat_front_addr":"192.168.123.109:6805/3122480642","state":["exists","up"]},{"osd":5,"uuid":"57003e78-f740-4009-98e4-a399b0d4d62a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":33,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6809","nonce":741270179}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6811","nonce":741270179}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6815","nonce":741270179}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6813","nonce":741270179}]},"public_addr":"192.168.123.109:6809/741270179","cluster_addr":"192.168.123.109:6811/741270179","heartbeat_back_addr":"192.168.123.109:6815/741270179","heartbeat_front_addr":"192.168.123.109:6813/741270179","state":["exists","up"]},{"osd":6,"uuid":"2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":39,"up_thru":40,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6817","nonce":536869529}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6819","nonce":536869529}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6823","nonce":536869529}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6821","nonce":536869529}]},"public_addr":"192.168.123.109:6817/536869529","cluster_addr":"192.168.123.109:6819/536869529","heartbeat_back_addr":"192.168.123.109:6823/536869529","heartbeat_front_addr":"192.168.123.109:6821/536869529","state":["exists","up"]},{"osd":7,"uuid":"8b7c23f9-10d4-4419-a925-798f85675eef","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":44,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6825","nonce":1888228353}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6827","nonce":1888228353}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6831","nonce":1888228353}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6829","nonce":1888228353}]},"public_addr":"192.168.123.109:6825/1888228353","cluster_addr":"192.168.123.109:6827/1888228353","heartbeat_back_addr":"192.168.123.109:6831/1888228353","heartbeat_front_addr":"192.168.123.109:6829/1888228353","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:30:46.209495+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:30:56.514069+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:05.898547+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:17.099125+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:26.176685+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:36.949367+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:48.497646+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[{"pgid":"1.0","osds":[0,6,1]}],"primary_temp":[],"blocklist":{"192.168.123.106:6801/2944919476":"2026-03-11T11:30:08.216815+0000","192.168.123.106:0/2364059127":"2026-03-11T11:30:08.216815+0000","192.168.123.106:6800/2944919476":"2026-03-11T11:30:08.216815+0000","192.168.123.106:0/3253507956":"2026-03-11T11:30:08.216815+0000","192.168.123.106:6801/2326359262":"2026-03-11T11:29:57.993690+0000","192.168.123.106:6800/2326359262":"2026-03-11T11:29:57.993690+0000","192.168.123.106:0/1447154569":"2026-03-11T11:29:57.993690+0000","192.168.123.106:0/1549600595":"2026-03-11T11:30:08.216815+0000","192.168.123.106:0/488585586":"2026-03-11T11:29:57.993690+0000","192.168.123.106:0/2251503666":"2026-03-11T11:29:57.993690+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[57405]: purged_snaps scrub starts 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[57405]: purged_snaps scrub ok 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[57405]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[57405]: osdmap e44: 8 total, 8 up, 8 in 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3286176118' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3323328979' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[49228]: purged_snaps scrub starts 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[49228]: purged_snaps scrub ok 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[49228]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[49228]: osdmap e44: 8 total, 8 up, 8 in 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3286176118' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T11:32:02.442 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:02 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3323328979' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:32:02.445 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-10T11:31:08.303126+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '19', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 7.889999866485596, 'score_stable': 7.889999866485596, 'optimal_score': 0.3799999952316284, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-10T11:32:02.445 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd pool get .mgr pg_num 2026-03-10T11:32:02.625 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:02.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:02 vm09 ceph-mon[54793]: purged_snaps scrub starts 2026-03-10T11:32:02.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:02 vm09 ceph-mon[54793]: purged_snaps scrub ok 2026-03-10T11:32:02.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:02 vm09 ceph-mon[54793]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:32:02.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:02 vm09 ceph-mon[54793]: osdmap e44: 8 total, 8 up, 8 in 2026-03-10T11:32:02.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:02 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:02.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:02 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3286176118' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T11:32:02.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:02 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3323328979' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:32:02.867 INFO:teuthology.orchestra.run.vm06.stdout:pg_num: 1 2026-03-10T11:32:02.936 INFO:tasks.cephadm:Adding ceph.rgw.foo.a on vm06 2026-03-10T11:32:02.936 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch apply rgw foo.a --placement '1;vm06=foo.a' 2026-03-10T11:32:03.115 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:32:03.375 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled rgw.foo.a update... 2026-03-10T11:32:03.427 DEBUG:teuthology.orchestra.run.vm06:rgw.foo.a> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@rgw.foo.a.service 2026-03-10T11:32:03.428 INFO:tasks.cephadm:Adding ceph.iscsi.iscsi.a on vm09 2026-03-10T11:32:03.428 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd pool create datapool 3 3 replicated 2026-03-10T11:32:03.616 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 434 MiB used, 160 GiB / 160 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: osdmap e45: 8 total, 8 up, 8 in 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2385769369' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.a", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.a", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:03.643 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:03 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 434 MiB used, 160 GiB / 160 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: osdmap e45: 8 total, 8 up, 8 in 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2385769369' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.a", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.a", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:03.691 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 434 MiB used, 160 GiB / 160 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: osdmap e45: 8 total, 8 up, 8 in 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2385769369' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.a", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.a", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:03.692 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:03 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: from='client.24319 -' entity='client.admin' cmd=[{"prefix": "orch apply rgw", "svc_id": "foo.a", "placement": "1;vm06=foo.a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: Saving service rgw.foo.a spec with placement vm06=foo.a;count:1 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: Deploying daemon rgw.foo.a on vm06 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: osdmap e46: 8 total, 8 up, 8 in 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/3676119558' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: from='client.24319 -' entity='client.admin' cmd=[{"prefix": "orch apply rgw", "svc_id": "foo.a", "placement": "1;vm06=foo.a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: Saving service rgw.foo.a spec with placement vm06=foo.a;count:1 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: Deploying daemon rgw.foo.a on vm06 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: osdmap e46: 8 total, 8 up, 8 in 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/3676119558' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T11:32:04.451 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.452 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.452 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.452 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.452 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.452 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:04 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:04.471 INFO:teuthology.orchestra.run.vm09.stderr:pool 'datapool' created 2026-03-10T11:32:04.524 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- rbd pool init datapool 2026-03-10T11:32:04.706 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: from='client.24319 -' entity='client.admin' cmd=[{"prefix": "orch apply rgw", "svc_id": "foo.a", "placement": "1;vm06=foo.a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: Saving service rgw.foo.a spec with placement vm06=foo.a;count:1 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: Deploying daemon rgw.foo.a on vm06 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: osdmap e46: 8 total, 8 up, 8 in 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/3676119558' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:04.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:04 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: Saving service rgw.foo.a spec with placement vm06=foo.a;count:1 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 433 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: osdmap e47: 8 total, 8 up, 8 in 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/4222316409' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/2480691246' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: Saving service rgw.foo.a spec with placement vm06=foo.a;count:1 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 433 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: osdmap e47: 8 total, 8 up, 8 in 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/4222316409' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/2480691246' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:05 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: Saving service rgw.foo.a spec with placement vm06=foo.a;count:1 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 433 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: osdmap e47: 8 total, 8 up, 8 in 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/4222316409' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/2480691246' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:05 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[49228]: Checking dashboard <-> RGW credentials 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[49228]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[49228]: osdmap e48: 8 total, 8 up, 8 in 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[49228]: osdmap e49: 8 total, 8 up, 8 in 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[49228]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[49228]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[57405]: Checking dashboard <-> RGW credentials 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[57405]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[57405]: osdmap e48: 8 total, 8 up, 8 in 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[57405]: osdmap e49: 8 total, 8 up, 8 in 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[57405]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:06.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:06 vm06 ceph-mon[57405]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:06 vm09 ceph-mon[54793]: Checking dashboard <-> RGW credentials 2026-03-10T11:32:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:06 vm09 ceph-mon[54793]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-10T11:32:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:06 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-10T11:32:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:06 vm09 ceph-mon[54793]: osdmap e48: 8 total, 8 up, 8 in 2026-03-10T11:32:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:06 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:06 vm09 ceph-mon[54793]: osdmap e49: 8 total, 8 up, 8 in 2026-03-10T11:32:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:06 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:06 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:06 vm09 ceph-mon[54793]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:06 vm09 ceph-mon[54793]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T11:32:07.533 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch apply iscsi datapool admin admin --trusted_ip_list 192.168.123.109 --placement '1;vm09=iscsi.a' 2026-03-10T11:32:07.715 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:32:07.747 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:07 vm09 ceph-mon[54793]: pgmap v97: 36 pgs: 35 unknown, 1 active+clean; 449 KiB data, 433 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:07.747 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:07 vm09 ceph-mon[54793]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-10T11:32:07.747 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:07 vm09 ceph-mon[54793]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-10T11:32:07.747 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:07 vm09 ceph-mon[54793]: osdmap e50: 8 total, 8 up, 8 in 2026-03-10T11:32:07.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:07 vm06 ceph-mon[57405]: pgmap v97: 36 pgs: 35 unknown, 1 active+clean; 449 KiB data, 433 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:07.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:07 vm06 ceph-mon[57405]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-10T11:32:07.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:07 vm06 ceph-mon[57405]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-10T11:32:07.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:07 vm06 ceph-mon[57405]: osdmap e50: 8 total, 8 up, 8 in 2026-03-10T11:32:07.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:07 vm06 ceph-mon[49228]: pgmap v97: 36 pgs: 35 unknown, 1 active+clean; 449 KiB data, 433 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:07.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:07 vm06 ceph-mon[49228]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-10T11:32:07.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:07 vm06 ceph-mon[49228]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-10T11:32:07.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:07 vm06 ceph-mon[49228]: osdmap e50: 8 total, 8 up, 8 in 2026-03-10T11:32:07.964 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled iscsi.datapool update... 2026-03-10T11:32:08.016 INFO:tasks.cephadm:Distributing iscsi-gateway.cfg... 2026-03-10T11:32:08.016 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:32:08.016 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/iscsi-gateway.cfg 2026-03-10T11:32:08.043 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:32:08.043 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/iscsi-gateway.cfg 2026-03-10T11:32:08.071 DEBUG:teuthology.orchestra.run.vm09:iscsi.iscsi.a> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@iscsi.iscsi.a.service 2026-03-10T11:32:08.114 INFO:tasks.cephadm:Adding prometheus.a on vm09 2026-03-10T11:32:08.114 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch apply prometheus '1;vm09=a' 2026-03-10T11:32:08.334 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:32:08.592 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled prometheus update... 2026-03-10T11:32:08.641 DEBUG:teuthology.orchestra.run.vm09:prometheus.a> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@prometheus.a.service 2026-03-10T11:32:08.642 INFO:tasks.cephadm:Adding node-exporter.a on vm06 2026-03-10T11:32:08.642 INFO:tasks.cephadm:Adding node-exporter.b on vm09 2026-03-10T11:32:08.642 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch apply node-exporter '2;vm06=a;vm09=b' 2026-03-10T11:32:08.859 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:32:09.115 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled node-exporter update... 2026-03-10T11:32:09.134 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:08 vm09 ceph-mon[54793]: from='client.24362 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.109", "placement": "1;vm09=iscsi.a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:09.134 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:08 vm09 ceph-mon[54793]: Saving service iscsi.datapool spec with placement vm09=iscsi.a;count:1 2026-03-10T11:32:09.134 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:08 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:09.135 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:08 vm09 ceph-mon[54793]: osdmap e51: 8 total, 8 up, 8 in 2026-03-10T11:32:09.135 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:08 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.135 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:08 vm09 ceph-mon[54793]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.135 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:08 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.135 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:08 vm09 ceph-mon[54793]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.135 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:08 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:09.164 DEBUG:teuthology.orchestra.run.vm06:node-exporter.a> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@node-exporter.a.service 2026-03-10T11:32:09.166 DEBUG:teuthology.orchestra.run.vm09:node-exporter.b> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@node-exporter.b.service 2026-03-10T11:32:09.168 INFO:tasks.cephadm:Adding alertmanager.a on vm06 2026-03-10T11:32:09.168 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch apply alertmanager '1;vm06=a' 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[49228]: from='client.24362 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.109", "placement": "1;vm09=iscsi.a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[49228]: Saving service iscsi.datapool spec with placement vm09=iscsi.a;count:1 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[49228]: osdmap e51: 8 total, 8 up, 8 in 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[49228]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[49228]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[57405]: from='client.24362 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.109", "placement": "1;vm09=iscsi.a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[57405]: Saving service iscsi.datapool spec with placement vm09=iscsi.a;count:1 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[57405]: osdmap e51: 8 total, 8 up, 8 in 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[57405]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[57405]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T11:32:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:08 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:09.392 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:32:09.670 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled alertmanager update... 2026-03-10T11:32:09.718 DEBUG:teuthology.orchestra.run.vm06:alertmanager.a> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@alertmanager.a.service 2026-03-10T11:32:09.720 INFO:tasks.cephadm:Adding grafana.a on vm09 2026-03-10T11:32:09.720 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph orch apply grafana '1;vm09=a' 2026-03-10T11:32:09.906 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:32:10.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[49228]: pgmap v100: 68 pgs: 3 creating+peering, 21 active+clean, 44 unknown; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 1023 B/s wr, 4 op/s 2026-03-10T11:32:10.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[49228]: from='client.24379 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "prometheus", "placement": "1;vm09=a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[49228]: Saving service prometheus spec with placement vm09=a;count:1 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[49228]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[49228]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[49228]: osdmap e52: 8 total, 8 up, 8 in 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[57405]: pgmap v100: 68 pgs: 3 creating+peering, 21 active+clean, 44 unknown; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 1023 B/s wr, 4 op/s 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[57405]: from='client.24379 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "prometheus", "placement": "1;vm09=a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[57405]: Saving service prometheus spec with placement vm09=a;count:1 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[57405]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[57405]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[57405]: osdmap e52: 8 total, 8 up, 8 in 2026-03-10T11:32:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:09 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:10.157 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled grafana update... 2026-03-10T11:32:10.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:09 vm09 ceph-mon[54793]: pgmap v100: 68 pgs: 3 creating+peering, 21 active+clean, 44 unknown; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 1023 B/s wr, 4 op/s 2026-03-10T11:32:10.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:09 vm09 ceph-mon[54793]: from='client.24379 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "prometheus", "placement": "1;vm09=a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:10.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:09 vm09 ceph-mon[54793]: Saving service prometheus spec with placement vm09=a;count:1 2026-03-10T11:32:10.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:09 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:10.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:09 vm09 ceph-mon[54793]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T11:32:10.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:09 vm09 ceph-mon[54793]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T11:32:10.197 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:09 vm09 ceph-mon[54793]: osdmap e52: 8 total, 8 up, 8 in 2026-03-10T11:32:10.198 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:09 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:10.221 DEBUG:teuthology.orchestra.run.vm09:grafana.a> sudo journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@grafana.a.service 2026-03-10T11:32:10.223 INFO:tasks.cephadm:Setting up client nodes... 2026-03-10T11:32:10.223 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-10T11:32:10.415 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:10.696 INFO:teuthology.orchestra.run.vm06.stdout:[client.0] 2026-03-10T11:32:10.696 INFO:teuthology.orchestra.run.vm06.stdout: key = AQA6AbBpPaEZKRAA2kl4ymp2ZWd3n1KszMBRDA== 2026-03-10T11:32:10.762 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:32:10.762 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-10T11:32:10.762 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-10T11:32:10.800 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph auth get-or-create client.1 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-10T11:32:10.982 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.b/config 2026-03-10T11:32:11.101 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: from='client.24385 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "node-exporter", "placement": "2;vm06=a;vm09=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:11.101 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: Saving service node-exporter spec with placement vm06=a;vm09=b;count:2 2026-03-10T11:32:11.101 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: from='client.24388 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "alertmanager", "placement": "1;vm06=a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:11.101 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: Saving service alertmanager spec with placement vm06=a;count:1 2026-03-10T11:32:11.102 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:11.102 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: osdmap e53: 8 total, 8 up, 8 in 2026-03-10T11:32:11.102 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.102 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.102 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.102 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.102 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/394480068' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T11:32:11.102 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:10 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/394480068' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T11:32:11.256 INFO:teuthology.orchestra.run.vm09.stdout:[client.1] 2026-03-10T11:32:11.256 INFO:teuthology.orchestra.run.vm09.stdout: key = AQA7AbBpyS4LDxAAHXZhLQGc6DpFTX+1oojPFg== 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: from='client.24385 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "node-exporter", "placement": "2;vm06=a;vm09=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: Saving service node-exporter spec with placement vm06=a;vm09=b;count:2 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: from='client.24388 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "alertmanager", "placement": "1;vm06=a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: Saving service alertmanager spec with placement vm06=a;count:1 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: osdmap e53: 8 total, 8 up, 8 in 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/394480068' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/394480068' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: from='client.24385 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "node-exporter", "placement": "2;vm06=a;vm09=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: Saving service node-exporter spec with placement vm06=a;vm09=b;count:2 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: from='client.24388 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "alertmanager", "placement": "1;vm06=a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: Saving service alertmanager spec with placement vm06=a;count:1 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: osdmap e53: 8 total, 8 up, 8 in 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/394480068' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T11:32:11.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:10 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/394480068' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T11:32:11.301 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T11:32:11.301 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.client.1.keyring 2026-03-10T11:32:11.301 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 0644 /etc/ceph/ceph.client.1.keyring 2026-03-10T11:32:11.346 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-10T11:32:11.346 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-10T11:32:11.347 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph mgr dump --format=json 2026-03-10T11:32:11.555 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:11.814 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:32:11.879 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":15,"flags":0,"active_gid":14150,"active_name":"y","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":649780914},{"type":"v1","addr":"192.168.123.106:6801","nonce":649780914}]},"active_addr":"192.168.123.106:6801/649780914","active_change":"2026-03-10T11:30:08.216899+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":24103,"name":"x","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.106:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":3312941136}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":2151786485}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":1008164583}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":3145469537}]}]} 2026-03-10T11:32:11.880 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-10T11:32:11.880 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-10T11:32:11.880 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd dump --format=json 2026-03-10T11:32:12.069 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:12.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: from='client.24380 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "grafana", "placement": "1;vm09=a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:12.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: Saving service grafana spec with placement vm09=a;count:1 2026-03-10T11:32:12.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: pgmap v103: 100 pgs: 20 creating+peering, 65 active+clean, 15 unknown; 450 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 3.5 KiB/s rd, 2.0 KiB/s wr, 6 op/s 2026-03-10T11:32:12.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/4017545829' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T11:32:12.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T11:32:12.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T11:32:12.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T11:32:12.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T11:32:12.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: osdmap e54: 8 total, 8 up, 8 in 2026-03-10T11:32:12.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.187 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1134364036' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: from='client.24380 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "grafana", "placement": "1;vm09=a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: Saving service grafana spec with placement vm09=a;count:1 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: pgmap v103: 100 pgs: 20 creating+peering, 65 active+clean, 15 unknown; 450 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 3.5 KiB/s rd, 2.0 KiB/s wr, 6 op/s 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/4017545829' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: osdmap e54: 8 total, 8 up, 8 in 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.188 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:11 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1134364036' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T11:32:12.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: from='client.24380 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "grafana", "placement": "1;vm09=a", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: Saving service grafana spec with placement vm09=a;count:1 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: pgmap v103: 100 pgs: 20 creating+peering, 65 active+clean, 15 unknown; 450 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 3.5 KiB/s rd, 2.0 KiB/s wr, 6 op/s 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/4017545829' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: osdmap e54: 8 total, 8 up, 8 in 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/966728943' entity='mgr.y' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/394226725' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: from='client.? ' entity='client.rgw.foo.a' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: from='client.? ' entity='mgr.y' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T11:32:12.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:11 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1134364036' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T11:32:12.308 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:32:12.308 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":54,"fsid":"52119272-1c74-11f1-990e-e7cc1c0e3258","created":"2026-03-10T11:29:47.393973+0000","modified":"2026-03-10T11:32:11.475994+0000","last_up_change":"2026-03-10T11:32:01.407381+0000","last_in_change":"2026-03-10T11:31:50.876994+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":6,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T11:31:08.303126+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"19","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"datapool","create_time":"2026-03-10T11:32:03.859891+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":3,"pg_placement_num":3,"pg_placement_num_target":3,"pg_num_target":3,"pg_num_pending":3,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"50","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":50,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2.6500000953674316,"score_stable":2.6500000953674316,"optimal_score":0.87999999523162842,"raw_score_acting":2.3299999237060547,"raw_score_stable":2.3299999237060547,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":3,"pool_name":".rgw.root","create_time":"2026-03-10T11:32:04.247540+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":32,"pg_placement_num":32,"pg_placement_num_target":32,"pg_num_target":32,"pg_num_pending":32,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"49","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rgw":{}},"read_balance":{"score_type":"Fair distribution","score_acting":1.5,"score_stable":1.5,"optimal_score":1,"raw_score_acting":1.5,"raw_score_stable":1.5,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":4,"pool_name":"default.rgw.log","create_time":"2026-03-10T11:32:05.578550+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":32,"pg_placement_num":32,"pg_placement_num_target":32,"pg_num_target":32,"pg_num_pending":32,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"51","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rgw":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2.25,"score_stable":2.25,"optimal_score":1,"raw_score_acting":2.25,"raw_score_stable":2.25,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":5,"pool_name":"default.rgw.control","create_time":"2026-03-10T11:32:07.496574+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":32,"pg_placement_num":32,"pg_placement_num_target":32,"pg_num_target":32,"pg_num_pending":32,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"53","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rgw":{}},"read_balance":{"score_type":"Fair distribution","score_acting":1.25,"score_stable":1.25,"optimal_score":1,"raw_score_acting":1.25,"raw_score_stable":1.25,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":6,"pool_name":"default.rgw.meta","create_time":"2026-03-10T11:32:09.610046+0000","flags":32769,"flags_names":"hashpspool,creating","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":32,"pg_placement_num":32,"pg_placement_num_target":32,"pg_num_target":32,"pg_num_pending":32,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"54","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rgw":{}},"read_balance":{"score_type":"Fair distribution","score_acting":1.75,"score_stable":1.75,"optimal_score":1,"raw_score_acting":1.75,"raw_score_stable":1.75,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"7704abb8-f49c-4aef-89e1-d9e67d3e765b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6803","nonce":3972625574}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6805","nonce":3972625574}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6809","nonce":3972625574}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6807","nonce":3972625574}]},"public_addr":"192.168.123.106:6803/3972625574","cluster_addr":"192.168.123.106:6805/3972625574","heartbeat_back_addr":"192.168.123.106:6809/3972625574","heartbeat_front_addr":"192.168.123.106:6807/3972625574","state":["exists","up"]},{"osd":1,"uuid":"f6896961-eee9-4f03-8e69-efccff3a919c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6811","nonce":193031054}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6813","nonce":193031054}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6817","nonce":193031054}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6815","nonce":193031054}]},"public_addr":"192.168.123.106:6811/193031054","cluster_addr":"192.168.123.106:6813/193031054","heartbeat_back_addr":"192.168.123.106:6817/193031054","heartbeat_front_addr":"192.168.123.106:6815/193031054","state":["exists","up"]},{"osd":2,"uuid":"600ee862-9ac1-423d-83b5-9b3e3288f4f1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6819","nonce":316237931}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6821","nonce":316237931}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6825","nonce":316237931}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6823","nonce":316237931}]},"public_addr":"192.168.123.106:6819/316237931","cluster_addr":"192.168.123.106:6821/316237931","heartbeat_back_addr":"192.168.123.106:6825/316237931","heartbeat_front_addr":"192.168.123.106:6823/316237931","state":["exists","up"]},{"osd":3,"uuid":"80cdadbb-309f-49c9-a457-c6534574dd5a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":22,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6827","nonce":1874161814}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6829","nonce":1874161814}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6832","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6833","nonce":1874161814}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6831","nonce":1874161814}]},"public_addr":"192.168.123.106:6827/1874161814","cluster_addr":"192.168.123.106:6829/1874161814","heartbeat_back_addr":"192.168.123.106:6833/1874161814","heartbeat_front_addr":"192.168.123.106:6831/1874161814","state":["exists","up"]},{"osd":4,"uuid":"3f0439f8-e455-4db3-88a6-7812e1b40aa7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6801","nonce":3122480642}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6803","nonce":3122480642}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6807","nonce":3122480642}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6805","nonce":3122480642}]},"public_addr":"192.168.123.109:6801/3122480642","cluster_addr":"192.168.123.109:6803/3122480642","heartbeat_back_addr":"192.168.123.109:6807/3122480642","heartbeat_front_addr":"192.168.123.109:6805/3122480642","state":["exists","up"]},{"osd":5,"uuid":"57003e78-f740-4009-98e4-a399b0d4d62a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6809","nonce":741270179}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6811","nonce":741270179}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6815","nonce":741270179}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6813","nonce":741270179}]},"public_addr":"192.168.123.109:6809/741270179","cluster_addr":"192.168.123.109:6811/741270179","heartbeat_back_addr":"192.168.123.109:6815/741270179","heartbeat_front_addr":"192.168.123.109:6813/741270179","state":["exists","up"]},{"osd":6,"uuid":"2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":39,"up_thru":51,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6817","nonce":536869529}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6819","nonce":536869529}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6823","nonce":536869529}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6821","nonce":536869529}]},"public_addr":"192.168.123.109:6817/536869529","cluster_addr":"192.168.123.109:6819/536869529","heartbeat_back_addr":"192.168.123.109:6823/536869529","heartbeat_front_addr":"192.168.123.109:6821/536869529","state":["exists","up"]},{"osd":7,"uuid":"8b7c23f9-10d4-4419-a925-798f85675eef","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":44,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6825","nonce":1888228353}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6827","nonce":1888228353}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6831","nonce":1888228353}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6829","nonce":1888228353}]},"public_addr":"192.168.123.109:6825/1888228353","cluster_addr":"192.168.123.109:6827/1888228353","heartbeat_back_addr":"192.168.123.109:6831/1888228353","heartbeat_front_addr":"192.168.123.109:6829/1888228353","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:30:46.209495+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:30:56.514069+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:05.898547+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:17.099125+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:26.176685+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:36.949367+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:48.497646+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:59.872145+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.106:6801/2944919476":"2026-03-11T11:30:08.216815+0000","192.168.123.106:0/2364059127":"2026-03-11T11:30:08.216815+0000","192.168.123.106:6800/2944919476":"2026-03-11T11:30:08.216815+0000","192.168.123.106:0/3253507956":"2026-03-11T11:30:08.216815+0000","192.168.123.106:6801/2326359262":"2026-03-11T11:29:57.993690+0000","192.168.123.106:6800/2326359262":"2026-03-11T11:29:57.993690+0000","192.168.123.106:0/1447154569":"2026-03-11T11:29:57.993690+0000","192.168.123.106:0/1549600595":"2026-03-11T11:30:08.216815+0000","192.168.123.106:0/488585586":"2026-03-11T11:29:57.993690+0000","192.168.123.106:0/2251503666":"2026-03-11T11:29:57.993690+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T11:32:12.378 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-10T11:32:12.379 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd dump --format=json 2026-03-10T11:32:12.614 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:12.939 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:32:12.939 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":55,"fsid":"52119272-1c74-11f1-990e-e7cc1c0e3258","created":"2026-03-10T11:29:47.393973+0000","modified":"2026-03-10T11:32:12.478863+0000","last_up_change":"2026-03-10T11:32:01.407381+0000","last_in_change":"2026-03-10T11:31:50.876994+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":6,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T11:31:08.303126+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"19","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"datapool","create_time":"2026-03-10T11:32:03.859891+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":3,"pg_placement_num":3,"pg_placement_num_target":3,"pg_num_target":3,"pg_num_pending":3,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"50","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":50,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2.6500000953674316,"score_stable":2.6500000953674316,"optimal_score":0.87999999523162842,"raw_score_acting":2.3299999237060547,"raw_score_stable":2.3299999237060547,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":3,"pool_name":".rgw.root","create_time":"2026-03-10T11:32:04.247540+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":32,"pg_placement_num":32,"pg_placement_num_target":32,"pg_num_target":32,"pg_num_pending":32,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"49","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rgw":{}},"read_balance":{"score_type":"Fair distribution","score_acting":1.5,"score_stable":1.5,"optimal_score":1,"raw_score_acting":1.5,"raw_score_stable":1.5,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":4,"pool_name":"default.rgw.log","create_time":"2026-03-10T11:32:05.578550+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":32,"pg_placement_num":32,"pg_placement_num_target":32,"pg_num_target":32,"pg_num_pending":32,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"51","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rgw":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2.25,"score_stable":2.25,"optimal_score":1,"raw_score_acting":2.25,"raw_score_stable":2.25,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":5,"pool_name":"default.rgw.control","create_time":"2026-03-10T11:32:07.496574+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":32,"pg_placement_num":32,"pg_placement_num_target":32,"pg_num_target":32,"pg_num_pending":32,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"53","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rgw":{}},"read_balance":{"score_type":"Fair distribution","score_acting":1.25,"score_stable":1.25,"optimal_score":1,"raw_score_acting":1.25,"raw_score_stable":1.25,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":6,"pool_name":"default.rgw.meta","create_time":"2026-03-10T11:32:09.610046+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":32,"pg_placement_num":32,"pg_placement_num_target":32,"pg_num_target":32,"pg_num_pending":32,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"55","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_autoscale_bias":4},"application_metadata":{"rgw":{}},"read_balance":{"score_type":"Fair distribution","score_acting":1.75,"score_stable":1.75,"optimal_score":1,"raw_score_acting":1.75,"raw_score_stable":1.75,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"7704abb8-f49c-4aef-89e1-d9e67d3e765b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6803","nonce":3972625574}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6805","nonce":3972625574}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6809","nonce":3972625574}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3972625574},{"type":"v1","addr":"192.168.123.106:6807","nonce":3972625574}]},"public_addr":"192.168.123.106:6803/3972625574","cluster_addr":"192.168.123.106:6805/3972625574","heartbeat_back_addr":"192.168.123.106:6809/3972625574","heartbeat_front_addr":"192.168.123.106:6807/3972625574","state":["exists","up"]},{"osd":1,"uuid":"f6896961-eee9-4f03-8e69-efccff3a919c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6811","nonce":193031054}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6813","nonce":193031054}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6817","nonce":193031054}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":193031054},{"type":"v1","addr":"192.168.123.106:6815","nonce":193031054}]},"public_addr":"192.168.123.106:6811/193031054","cluster_addr":"192.168.123.106:6813/193031054","heartbeat_back_addr":"192.168.123.106:6817/193031054","heartbeat_front_addr":"192.168.123.106:6815/193031054","state":["exists","up"]},{"osd":2,"uuid":"600ee862-9ac1-423d-83b5-9b3e3288f4f1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6819","nonce":316237931}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6821","nonce":316237931}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6825","nonce":316237931}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":316237931},{"type":"v1","addr":"192.168.123.106:6823","nonce":316237931}]},"public_addr":"192.168.123.106:6819/316237931","cluster_addr":"192.168.123.106:6821/316237931","heartbeat_back_addr":"192.168.123.106:6825/316237931","heartbeat_front_addr":"192.168.123.106:6823/316237931","state":["exists","up"]},{"osd":3,"uuid":"80cdadbb-309f-49c9-a457-c6534574dd5a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":22,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6827","nonce":1874161814}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6829","nonce":1874161814}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6832","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6833","nonce":1874161814}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":1874161814},{"type":"v1","addr":"192.168.123.106:6831","nonce":1874161814}]},"public_addr":"192.168.123.106:6827/1874161814","cluster_addr":"192.168.123.106:6829/1874161814","heartbeat_back_addr":"192.168.123.106:6833/1874161814","heartbeat_front_addr":"192.168.123.106:6831/1874161814","state":["exists","up"]},{"osd":4,"uuid":"3f0439f8-e455-4db3-88a6-7812e1b40aa7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6801","nonce":3122480642}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6803","nonce":3122480642}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6807","nonce":3122480642}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":3122480642},{"type":"v1","addr":"192.168.123.109:6805","nonce":3122480642}]},"public_addr":"192.168.123.109:6801/3122480642","cluster_addr":"192.168.123.109:6803/3122480642","heartbeat_back_addr":"192.168.123.109:6807/3122480642","heartbeat_front_addr":"192.168.123.109:6805/3122480642","state":["exists","up"]},{"osd":5,"uuid":"57003e78-f740-4009-98e4-a399b0d4d62a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6809","nonce":741270179}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6811","nonce":741270179}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6815","nonce":741270179}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":741270179},{"type":"v1","addr":"192.168.123.109:6813","nonce":741270179}]},"public_addr":"192.168.123.109:6809/741270179","cluster_addr":"192.168.123.109:6811/741270179","heartbeat_back_addr":"192.168.123.109:6815/741270179","heartbeat_front_addr":"192.168.123.109:6813/741270179","state":["exists","up"]},{"osd":6,"uuid":"2a7c5fa5-1b06-4707-bf6d-ace5f9fcc111","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":39,"up_thru":51,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6817","nonce":536869529}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6819","nonce":536869529}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6823","nonce":536869529}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":536869529},{"type":"v1","addr":"192.168.123.109:6821","nonce":536869529}]},"public_addr":"192.168.123.109:6817/536869529","cluster_addr":"192.168.123.109:6819/536869529","heartbeat_back_addr":"192.168.123.109:6823/536869529","heartbeat_front_addr":"192.168.123.109:6821/536869529","state":["exists","up"]},{"osd":7,"uuid":"8b7c23f9-10d4-4419-a925-798f85675eef","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":44,"up_thru":53,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6825","nonce":1888228353}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6827","nonce":1888228353}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6831","nonce":1888228353}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":1888228353},{"type":"v1","addr":"192.168.123.109:6829","nonce":1888228353}]},"public_addr":"192.168.123.109:6825/1888228353","cluster_addr":"192.168.123.109:6827/1888228353","heartbeat_back_addr":"192.168.123.109:6831/1888228353","heartbeat_front_addr":"192.168.123.109:6829/1888228353","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:30:46.209495+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:30:56.514069+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:05.898547+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:17.099125+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:26.176685+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:36.949367+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:48.497646+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T11:31:59.872145+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.106:6801/2944919476":"2026-03-11T11:30:08.216815+0000","192.168.123.106:0/2364059127":"2026-03-11T11:30:08.216815+0000","192.168.123.106:6800/2944919476":"2026-03-11T11:30:08.216815+0000","192.168.123.106:0/3253507956":"2026-03-11T11:30:08.216815+0000","192.168.123.106:6801/2326359262":"2026-03-11T11:29:57.993690+0000","192.168.123.106:6800/2326359262":"2026-03-11T11:29:57.993690+0000","192.168.123.106:0/1447154569":"2026-03-11T11:29:57.993690+0000","192.168.123.106:0/1549600595":"2026-03-11T11:30:08.216815+0000","192.168.123.106:0/488585586":"2026-03-11T11:29:57.993690+0000","192.168.123.106:0/2251503666":"2026-03-11T11:29:57.993690+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T11:32:12.992 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph tell osd.0 flush_pg_stats 2026-03-10T11:32:12.993 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph tell osd.1 flush_pg_stats 2026-03-10T11:32:12.993 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph tell osd.2 flush_pg_stats 2026-03-10T11:32:12.993 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph tell osd.3 flush_pg_stats 2026-03-10T11:32:12.993 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph tell osd.4 flush_pg_stats 2026-03-10T11:32:12.993 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph tell osd.5 flush_pg_stats 2026-03-10T11:32:12.993 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph tell osd.6 flush_pg_stats 2026-03-10T11:32:12.994 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph tell osd.7 flush_pg_stats 2026-03-10T11:32:13.194 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[49228]: pgmap v106: 132 pgs: 23 creating+peering, 75 active+clean, 34 unknown; 450 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1.2 KiB/s wr, 4 op/s 2026-03-10T11:32:13.194 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3787166670' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:32:13.194 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[49228]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T11:32:13.194 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[49228]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[49228]: osdmap e55: 8 total, 8 up, 8 in 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3037583315' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[57405]: pgmap v106: 132 pgs: 23 creating+peering, 75 active+clean, 34 unknown; 450 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1.2 KiB/s wr, 4 op/s 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3787166670' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[57405]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[57405]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[57405]: osdmap e55: 8 total, 8 up, 8 in 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3037583315' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:13.195 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:12 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:13.284 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:12 vm09 ceph-mon[54793]: pgmap v106: 132 pgs: 23 creating+peering, 75 active+clean, 34 unknown; 450 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1.2 KiB/s wr, 4 op/s 2026-03-10T11:32:13.284 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:12 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3787166670' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:32:13.284 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:12 vm09 ceph-mon[54793]: from='client.? ' entity='client.rgw.foo.a' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T11:32:13.284 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:12 vm09 ceph-mon[54793]: from='client.? ' entity='mgr.y' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T11:32:13.284 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:12 vm09 ceph-mon[54793]: osdmap e55: 8 total, 8 up, 8 in 2026-03-10T11:32:13.284 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:12 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:13.284 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:12 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3037583315' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:32:13.284 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:12 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:13.284 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:12 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:13.849 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:13.859 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:13 vm09 systemd[1]: Starting Ceph iscsi.iscsi.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:32:13.864 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:13.896 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:13.945 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:13.971 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:14.112 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:13 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-10T11:32:14.112 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:13 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-10T11:32:14.112 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:13 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:14.112 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:13 vm09 ceph-mon[54793]: Deploying daemon iscsi.iscsi.a on vm09 2026-03-10T11:32:14.112 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:13 vm09 ceph-mon[54793]: osdmap e56: 8 total, 8 up, 8 in 2026-03-10T11:32:14.112 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:13 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.112 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:13 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.112 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:13 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.112 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:13 vm09 ceph-mon[54793]: Checking pool "datapool" exists for service iscsi.datapool 2026-03-10T11:32:14.112 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:13 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.112 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:13 vm09 podman[78944]: 2026-03-10 11:32:13.859498424 +0000 UTC m=+0.018529815 container create 42354789c474e6bce60aeb219bc9d5059ca7937e6e9960ef4a0a84851a99b781 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, io.buildah.version=1.41.3, ceph=True, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2) 2026-03-10T11:32:14.112 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:13 vm09 podman[78944]: 2026-03-10 11:32:13.901509468 +0000 UTC m=+0.060540859 container init 42354789c474e6bce60aeb219bc9d5059ca7937e6e9960ef4a0a84851a99b781 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.license=GPLv2, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223) 2026-03-10T11:32:14.112 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:13 vm09 podman[78944]: 2026-03-10 11:32:13.905464641 +0000 UTC m=+0.064496032 container start 42354789c474e6bce60aeb219bc9d5059ca7937e6e9960ef4a0a84851a99b781 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.label-schema.schema-version=1.0, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-10T11:32:14.113 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:13 vm09 bash[78944]: 42354789c474e6bce60aeb219bc9d5059ca7937e6e9960ef4a0a84851a99b781 2026-03-10T11:32:14.113 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:13 vm09 podman[78944]: 2026-03-10 11:32:13.852228064 +0000 UTC m=+0.011259456 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T11:32:14.113 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:13 vm09 systemd[1]: Started Ceph iscsi.iscsi.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[49228]: Deploying daemon iscsi.iscsi.a on vm09 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[49228]: osdmap e56: 8 total, 8 up, 8 in 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[49228]: Checking pool "datapool" exists for service iscsi.datapool 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[57405]: Deploying daemon iscsi.iscsi.a on vm09 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[57405]: osdmap e56: 8 total, 8 up, 8 in 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[57405]: Checking pool "datapool" exists for service iscsi.datapool 2026-03-10T11:32:14.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:13 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:14.344 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:14.348 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:14.355 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:14.366 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug Started the configuration object watcher 2026-03-10T11:32:14.366 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug Checking for config object changes every 1s 2026-03-10T11:32:14.366 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug Processing osd blocklist entries for this node 2026-03-10T11:32:14.366 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug Reading the configuration object to update local LIO configuration 2026-03-10T11:32:14.366 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug Configuration does not have an entry for this host(vm09.local) - nothing to define to LIO 2026-03-10T11:32:14.367 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: * Serving Flask app 'rbd-target-api' (lazy loading) 2026-03-10T11:32:14.367 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: * Environment: production 2026-03-10T11:32:14.367 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-10T11:32:14.367 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: Use a production WSGI server instead. 2026-03-10T11:32:14.367 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: * Debug mode: off 2026-03-10T11:32:14.367 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug * Running on all addresses. 2026-03-10T11:32:14.367 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-10T11:32:14.367 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: * Running on all addresses. 2026-03-10T11:32:14.367 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-10T11:32:14.367 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug * Running on http://[::1]:5000/ (Press CTRL+C to quit) 2026-03-10T11:32:14.367 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:14 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: * Running on http://[::1]:5000/ (Press CTRL+C to quit) 2026-03-10T11:32:14.937 INFO:teuthology.orchestra.run.vm06.stdout:51539607569 2026-03-10T11:32:14.937 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.1 2026-03-10T11:32:14.979 INFO:teuthology.orchestra.run.vm06.stdout:167503724550 2026-03-10T11:32:14.979 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.6 2026-03-10T11:32:14.991 INFO:teuthology.orchestra.run.vm06.stdout:115964117003 2026-03-10T11:32:14.991 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.4 2026-03-10T11:32:15.059 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:15 vm06 ceph-mon[49228]: pgmap v109: 132 pgs: 6 creating+peering, 126 active+clean; 454 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 95 KiB/s rd, 8.2 KiB/s wr, 232 op/s 2026-03-10T11:32:15.059 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:15 vm06 ceph-mon[49228]: Deploying daemon prometheus.a on vm09 2026-03-10T11:32:15.059 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:15 vm06 ceph-mon[49228]: from='client.? 192.168.123.109:0/355638427' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-10T11:32:15.059 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:15 vm06 ceph-mon[57405]: pgmap v109: 132 pgs: 6 creating+peering, 126 active+clean; 454 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 95 KiB/s rd, 8.2 KiB/s wr, 232 op/s 2026-03-10T11:32:15.059 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:15 vm06 ceph-mon[57405]: Deploying daemon prometheus.a on vm09 2026-03-10T11:32:15.059 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:15 vm06 ceph-mon[57405]: from='client.? 192.168.123.109:0/355638427' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-10T11:32:15.087 INFO:teuthology.orchestra.run.vm06.stdout:68719476751 2026-03-10T11:32:15.087 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.2 2026-03-10T11:32:15.088 INFO:teuthology.orchestra.run.vm06.stdout:94489280525 2026-03-10T11:32:15.088 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.3 2026-03-10T11:32:15.100 INFO:teuthology.orchestra.run.vm06.stdout:137438953481 2026-03-10T11:32:15.100 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.5 2026-03-10T11:32:15.121 INFO:teuthology.orchestra.run.vm06.stdout:34359738387 2026-03-10T11:32:15.121 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.0 2026-03-10T11:32:15.222 INFO:teuthology.orchestra.run.vm06.stdout:188978561028 2026-03-10T11:32:15.222 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.7 2026-03-10T11:32:15.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:15 vm09 ceph-mon[54793]: pgmap v109: 132 pgs: 6 creating+peering, 126 active+clean; 454 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 95 KiB/s rd, 8.2 KiB/s wr, 232 op/s 2026-03-10T11:32:15.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:15 vm09 ceph-mon[54793]: Deploying daemon prometheus.a on vm09 2026-03-10T11:32:15.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:15 vm09 ceph-mon[54793]: from='client.? 192.168.123.109:0/355638427' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-10T11:32:15.624 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:15.694 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:15.882 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:16.158 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:16.191 INFO:teuthology.orchestra.run.vm06.stdout:51539607568 2026-03-10T11:32:16.217 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:16.242 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:16.250 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:16 vm06 ceph-mon[49228]: mgrmap e16: y(active, since 2m), standbys: x 2026-03-10T11:32:16.250 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:16 vm06 ceph-mon[57405]: mgrmap e16: y(active, since 2m), standbys: x 2026-03-10T11:32:16.428 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607569 got 51539607568 for osd.1 2026-03-10T11:32:16.438 INFO:teuthology.orchestra.run.vm06.stdout:115964117002 2026-03-10T11:32:16.452 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:16.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:16 vm09 ceph-mon[54793]: mgrmap e16: y(active, since 2m), standbys: x 2026-03-10T11:32:16.481 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:16.601 INFO:teuthology.orchestra.run.vm06.stdout:68719476750 2026-03-10T11:32:16.611 INFO:tasks.cephadm.ceph_manager.ceph:need seq 115964117003 got 115964117002 for osd.4 2026-03-10T11:32:16.756 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476751 got 68719476750 for osd.2 2026-03-10T11:32:16.986 INFO:teuthology.orchestra.run.vm06.stdout:137438953480 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3363722398' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[49228]: pgmap v110: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 80 KiB/s rd, 6.4 KiB/s wr, 196 op/s 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/902170765' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3178136051' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3866792668' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3363722398' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[57405]: pgmap v110: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 80 KiB/s rd, 6.4 KiB/s wr, 196 op/s 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/902170765' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3178136051' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T11:32:17.042 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:17 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3866792668' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T11:32:17.102 INFO:tasks.cephadm.ceph_manager.ceph:need seq 137438953481 got 137438953480 for osd.5 2026-03-10T11:32:17.112 INFO:teuthology.orchestra.run.vm06.stdout:188978561027 2026-03-10T11:32:17.139 INFO:teuthology.orchestra.run.vm06.stdout:94489280524 2026-03-10T11:32:17.189 INFO:tasks.cephadm.ceph_manager.ceph:need seq 188978561028 got 188978561027 for osd.7 2026-03-10T11:32:17.189 INFO:teuthology.orchestra.run.vm06.stdout:34359738386 2026-03-10T11:32:17.248 INFO:tasks.cephadm.ceph_manager.ceph:need seq 94489280525 got 94489280524 for osd.3 2026-03-10T11:32:17.249 INFO:teuthology.orchestra.run.vm06.stdout:167503724549 2026-03-10T11:32:17.312 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738387 got 34359738386 for osd.0 2026-03-10T11:32:17.357 INFO:tasks.cephadm.ceph_manager.ceph:need seq 167503724550 got 167503724549 for osd.6 2026-03-10T11:32:17.428 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.1 2026-03-10T11:32:17.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:17 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3363722398' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T11:32:17.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:17 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:17.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:17 vm09 ceph-mon[54793]: pgmap v110: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 80 KiB/s rd, 6.4 KiB/s wr, 196 op/s 2026-03-10T11:32:17.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:17 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/902170765' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T11:32:17.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:17 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3178136051' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T11:32:17.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:17 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3866792668' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T11:32:17.612 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.4 2026-03-10T11:32:17.615 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:17.757 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.2 2026-03-10T11:32:17.888 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:17.984 INFO:teuthology.orchestra.run.vm06.stdout:51539607569 2026-03-10T11:32:18.067 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:18.085 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607569 got 51539607569 for osd.1 2026-03-10T11:32:18.085 DEBUG:teuthology.parallel:result is None 2026-03-10T11:32:18.103 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.5 2026-03-10T11:32:18.189 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.7 2026-03-10T11:32:18.241 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:18 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/562145878' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T11:32:18.241 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:18 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2624051469' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T11:32:18.241 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:18 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2780333570' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T11:32:18.242 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:18 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2685562401' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T11:32:18.242 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:18 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2532611111' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T11:32:18.242 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:18 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/562145878' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T11:32:18.242 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:18 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2624051469' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T11:32:18.242 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:18 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2780333570' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T11:32:18.242 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:18 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2685562401' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T11:32:18.242 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:18 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2532611111' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T11:32:18.249 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.3 2026-03-10T11:32:18.313 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.0 2026-03-10T11:32:18.358 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph osd last-stat-seq osd.6 2026-03-10T11:32:18.374 INFO:teuthology.orchestra.run.vm06.stdout:115964117003 2026-03-10T11:32:18.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:18 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/562145878' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T11:32:18.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:18 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2624051469' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T11:32:18.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:18 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2780333570' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T11:32:18.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:18 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2685562401' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T11:32:18.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:18 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2532611111' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T11:32:18.538 INFO:teuthology.orchestra.run.vm06.stdout:68719476751 2026-03-10T11:32:18.590 INFO:tasks.cephadm.ceph_manager.ceph:need seq 115964117003 got 115964117003 for osd.4 2026-03-10T11:32:18.591 DEBUG:teuthology.parallel:result is None 2026-03-10T11:32:18.633 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:18.670 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476751 got 68719476751 for osd.2 2026-03-10T11:32:18.670 DEBUG:teuthology.parallel:result is None 2026-03-10T11:32:18.781 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:18.995 INFO:teuthology.orchestra.run.vm06.stdout:137438953481 2026-03-10T11:32:19.065 INFO:tasks.cephadm.ceph_manager.ceph:need seq 137438953481 got 137438953481 for osd.5 2026-03-10T11:32:19.065 DEBUG:teuthology.parallel:result is None 2026-03-10T11:32:19.169 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:19.170 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:19.172 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:19.233 INFO:teuthology.orchestra.run.vm06.stdout:188978561029 2026-03-10T11:32:19.248 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:19 vm06 ceph-mon[49228]: pgmap v111: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 69 KiB/s rd, 5.4 KiB/s wr, 167 op/s 2026-03-10T11:32:19.248 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:19 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3734954664' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T11:32:19.248 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:19 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/195767662' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T11:32:19.248 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:19 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2054582527' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T11:32:19.249 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:19 vm06 ceph-mon[57405]: pgmap v111: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 69 KiB/s rd, 5.4 KiB/s wr, 167 op/s 2026-03-10T11:32:19.249 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:19 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3734954664' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T11:32:19.249 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:19 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/195767662' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T11:32:19.249 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:19 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2054582527' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T11:32:19.356 INFO:tasks.cephadm.ceph_manager.ceph:need seq 188978561028 got 188978561029 for osd.7 2026-03-10T11:32:19.356 DEBUG:teuthology.parallel:result is None 2026-03-10T11:32:19.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:19 vm09 ceph-mon[54793]: pgmap v111: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 69 KiB/s rd, 5.4 KiB/s wr, 167 op/s 2026-03-10T11:32:19.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3734954664' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T11:32:19.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/195767662' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T11:32:19.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2054582527' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T11:32:19.610 INFO:teuthology.orchestra.run.vm06.stdout:167503724551 2026-03-10T11:32:19.625 INFO:teuthology.orchestra.run.vm06.stdout:34359738388 2026-03-10T11:32:19.695 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738387 got 34359738388 for osd.0 2026-03-10T11:32:19.695 DEBUG:teuthology.parallel:result is None 2026-03-10T11:32:19.701 INFO:teuthology.orchestra.run.vm06.stdout:94489280525 2026-03-10T11:32:19.762 INFO:tasks.cephadm.ceph_manager.ceph:need seq 167503724550 got 167503724551 for osd.6 2026-03-10T11:32:19.762 DEBUG:teuthology.parallel:result is None 2026-03-10T11:32:19.764 INFO:tasks.cephadm.ceph_manager.ceph:need seq 94489280525 got 94489280525 for osd.3 2026-03-10T11:32:19.764 DEBUG:teuthology.parallel:result is None 2026-03-10T11:32:19.764 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-10T11:32:19.764 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph pg dump --format=json 2026-03-10T11:32:19.981 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:20.146 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:20 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3456264081' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T11:32:20.146 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:20 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3780392221' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T11:32:20.146 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:20 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2502318196' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T11:32:20.146 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:20 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3541135833' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T11:32:20.146 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 systemd[1]: Starting Ceph prometheus.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:32:20.229 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:32:20.232 INFO:teuthology.orchestra.run.vm06.stderr:dumped all 2026-03-10T11:32:20.262 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:20 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3456264081' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T11:32:20.263 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:20 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3780392221' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T11:32:20.263 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:20 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2502318196' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T11:32:20.263 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:20 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3541135833' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T11:32:20.263 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3456264081' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T11:32:20.263 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3780392221' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T11:32:20.263 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2502318196' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T11:32:20.263 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3541135833' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T11:32:20.288 INFO:teuthology.orchestra.run.vm06.stdout:{"pg_ready":true,"pg_map":{"version":111,"stamp":"2026-03-10T11:32:18.277231+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":465419,"num_objects":199,"num_object_clones":0,"num_object_copies":597,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":199,"num_whiteouts":0,"num_read":776,"num_read_kb":519,"num_write":493,"num_write_kb":629,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":459280,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":505,"ondisk_log_size":505,"up":396,"acting":396,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":396,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":8,"kb":167739392,"kb_used":221028,"kb_used_data":6316,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167518364,"statfs":{"total":171765137408,"available":171538804736,"internally_reserved":0,"allocated":6467584,"data_stored":3447344,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12711,"internal_metadata":219663961},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":33,"apply_latency_ms":33,"commit_latency_ns":33000000,"apply_latency_ns":33000000},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":4507,"num_objects":188,"num_object_clones":0,"num_object_copies":564,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":188,"num_whiteouts":0,"num_read":716,"num_read_kb":468,"num_write":424,"num_write_kb":37,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"6.796821"},"pg_stats":[{"pgid":"3.1f","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.538855+0000","last_change":"2026-03-10T11:32:05.482279+0000","last_active":"2026-03-10T11:32:13.538855+0000","last_peered":"2026-03-10T11:32:13.538855+0000","last_clean":"2026-03-10T11:32:13.538855+0000","last_became_active":"2026-03-10T11:32:05.482189+0000","last_became_peered":"2026-03-10T11:32:05.482189+0000","last_unstale":"2026-03-10T11:32:13.538855+0000","last_undegraded":"2026-03-10T11:32:13.538855+0000","last_fullsized":"2026-03-10T11:32:13.538855+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:33:08.999865+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,5,2],"acting":[0,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.18","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059914+0000","last_change":"2026-03-10T11:32:07.478618+0000","last_active":"2026-03-10T11:32:14.059914+0000","last_peered":"2026-03-10T11:32:14.059914+0000","last_clean":"2026-03-10T11:32:14.059914+0000","last_became_active":"2026-03-10T11:32:07.478521+0000","last_became_peered":"2026-03-10T11:32:07.478521+0000","last_unstale":"2026-03-10T11:32:14.059914+0000","last_undegraded":"2026-03-10T11:32:14.059914+0000","last_fullsized":"2026-03-10T11:32:14.059914+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:31:56.427965+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,6,1],"acting":[4,6,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"5.19","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757486+0000","last_change":"2026-03-10T11:32:09.511642+0000","last_active":"2026-03-10T11:32:13.757486+0000","last_peered":"2026-03-10T11:32:13.757486+0000","last_clean":"2026-03-10T11:32:13.757486+0000","last_became_active":"2026-03-10T11:32:09.509697+0000","last_became_peered":"2026-03-10T11:32:09.509697+0000","last_unstale":"2026-03-10T11:32:13.757486+0000","last_undegraded":"2026-03-10T11:32:13.757486+0000","last_fullsized":"2026-03-10T11:32:13.757486+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:00:25.667302+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,7],"acting":[1,5,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"6.1a","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059887+0000","last_change":"2026-03-10T11:32:11.509684+0000","last_active":"2026-03-10T11:32:14.059887+0000","last_peered":"2026-03-10T11:32:14.059887+0000","last_clean":"2026-03-10T11:32:14.059887+0000","last_became_active":"2026-03-10T11:32:11.509561+0000","last_became_peered":"2026-03-10T11:32:11.509561+0000","last_unstale":"2026-03-10T11:32:14.059887+0000","last_undegraded":"2026-03-10T11:32:14.059887+0000","last_fullsized":"2026-03-10T11:32:14.059887+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:03:55.104524+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,5,1],"acting":[4,5,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"6.1b","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432284+0000","last_change":"2026-03-10T11:32:11.573329+0000","last_active":"2026-03-10T11:32:13.432284+0000","last_peered":"2026-03-10T11:32:13.432284+0000","last_clean":"2026-03-10T11:32:13.432284+0000","last_became_active":"2026-03-10T11:32:11.573182+0000","last_became_peered":"2026-03-10T11:32:11.573182+0000","last_unstale":"2026-03-10T11:32:13.432284+0000","last_undegraded":"2026-03-10T11:32:13.432284+0000","last_fullsized":"2026-03-10T11:32:13.432284+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:50:42.681652+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,6],"acting":[3,7,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.1e","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432298+0000","last_change":"2026-03-10T11:32:05.482297+0000","last_active":"2026-03-10T11:32:13.432298+0000","last_peered":"2026-03-10T11:32:13.432298+0000","last_clean":"2026-03-10T11:32:13.432298+0000","last_became_active":"2026-03-10T11:32:05.482171+0000","last_became_peered":"2026-03-10T11:32:05.482171+0000","last_unstale":"2026-03-10T11:32:13.432298+0000","last_undegraded":"2026-03-10T11:32:13.432298+0000","last_fullsized":"2026-03-10T11:32:13.432298+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:02:00.446821+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,6,2],"acting":[3,6,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.19","version":"55'15","reported_seq":48,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432324+0000","last_change":"2026-03-10T11:32:07.482046+0000","last_active":"2026-03-10T11:32:13.432324+0000","last_peered":"2026-03-10T11:32:13.432324+0000","last_clean":"2026-03-10T11:32:13.432324+0000","last_became_active":"2026-03-10T11:32:07.481957+0000","last_became_peered":"2026-03-10T11:32:07.481957+0000","last_unstale":"2026-03-10T11:32:13.432324+0000","last_undegraded":"2026-03-10T11:32:13.432324+0000","last_fullsized":"2026-03-10T11:32:13.432324+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":15,"log_dups_size":0,"ondisk_log_size":15,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:56:19.989002+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":7,"num_object_clones":0,"num_object_copies":21,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":7,"num_whiteouts":0,"num_read":26,"num_read_kb":17,"num_write":16,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,2,0],"acting":[3,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.18","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058627+0000","last_change":"2026-03-10T11:32:09.514728+0000","last_active":"2026-03-10T11:32:14.058627+0000","last_peered":"2026-03-10T11:32:14.058627+0000","last_clean":"2026-03-10T11:32:14.058627+0000","last_became_active":"2026-03-10T11:32:09.509041+0000","last_became_peered":"2026-03-10T11:32:09.509041+0000","last_unstale":"2026-03-10T11:32:14.058627+0000","last_undegraded":"2026-03-10T11:32:14.058627+0000","last_fullsized":"2026-03-10T11:32:14.058627+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:23:50.009456+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,6,1],"acting":[4,6,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.1d","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434914+0000","last_change":"2026-03-10T11:32:05.477572+0000","last_active":"2026-03-10T11:32:13.434914+0000","last_peered":"2026-03-10T11:32:13.434914+0000","last_clean":"2026-03-10T11:32:13.434914+0000","last_became_active":"2026-03-10T11:32:05.477441+0000","last_became_peered":"2026-03-10T11:32:05.477441+0000","last_unstale":"2026-03-10T11:32:13.434914+0000","last_undegraded":"2026-03-10T11:32:13.434914+0000","last_fullsized":"2026-03-10T11:32:13.434914+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:49:49.474258+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,4,6],"acting":[5,4,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"4.1a","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058913+0000","last_change":"2026-03-10T11:32:07.480994+0000","last_active":"2026-03-10T11:32:14.058913+0000","last_peered":"2026-03-10T11:32:14.058913+0000","last_clean":"2026-03-10T11:32:14.058913+0000","last_became_active":"2026-03-10T11:32:07.480797+0000","last_became_peered":"2026-03-10T11:32:07.480797+0000","last_unstale":"2026-03-10T11:32:14.058913+0000","last_undegraded":"2026-03-10T11:32:14.058913+0000","last_fullsized":"2026-03-10T11:32:14.058913+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:08:46.328816+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,0],"acting":[4,3,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"5.1b","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434880+0000","last_change":"2026-03-10T11:32:09.513311+0000","last_active":"2026-03-10T11:32:13.434880+0000","last_peered":"2026-03-10T11:32:13.434880+0000","last_clean":"2026-03-10T11:32:13.434880+0000","last_became_active":"2026-03-10T11:32:09.513156+0000","last_became_peered":"2026-03-10T11:32:09.513156+0000","last_unstale":"2026-03-10T11:32:13.434880+0000","last_undegraded":"2026-03-10T11:32:13.434880+0000","last_fullsized":"2026-03-10T11:32:13.434880+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:07:26.592760+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,0,7],"acting":[5,0,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.18","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539342+0000","last_change":"2026-03-10T11:32:11.512406+0000","last_active":"2026-03-10T11:32:13.539342+0000","last_peered":"2026-03-10T11:32:13.539342+0000","last_clean":"2026-03-10T11:32:13.539342+0000","last_became_active":"2026-03-10T11:32:11.512324+0000","last_became_peered":"2026-03-10T11:32:11.512324+0000","last_unstale":"2026-03-10T11:32:13.539342+0000","last_undegraded":"2026-03-10T11:32:13.539342+0000","last_fullsized":"2026-03-10T11:32:13.539342+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:47:15.405374+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,1,7],"acting":[0,1,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"3.1c","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435186+0000","last_change":"2026-03-10T11:32:05.477512+0000","last_active":"2026-03-10T11:32:13.435186+0000","last_peered":"2026-03-10T11:32:13.435186+0000","last_clean":"2026-03-10T11:32:13.435186+0000","last_became_active":"2026-03-10T11:32:05.477337+0000","last_became_peered":"2026-03-10T11:32:05.477337+0000","last_unstale":"2026-03-10T11:32:13.435186+0000","last_undegraded":"2026-03-10T11:32:13.435186+0000","last_fullsized":"2026-03-10T11:32:13.435186+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:38:48.039126+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,4,1],"acting":[5,4,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"4.1b","version":"55'5","reported_seq":33,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058878+0000","last_change":"2026-03-10T11:32:07.483080+0000","last_active":"2026-03-10T11:32:14.058878+0000","last_peered":"2026-03-10T11:32:14.058878+0000","last_clean":"2026-03-10T11:32:14.058878+0000","last_became_active":"2026-03-10T11:32:07.482661+0000","last_became_peered":"2026-03-10T11:32:07.482661+0000","last_unstale":"2026-03-10T11:32:14.058878+0000","last_undegraded":"2026-03-10T11:32:14.058878+0000","last_fullsized":"2026-03-10T11:32:14.058878+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":5,"log_dups_size":0,"ondisk_log_size":5,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:56:09.413634+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":11,"num_read_kb":7,"num_write":6,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,1],"acting":[4,3,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"5.1a","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752686+0000","last_change":"2026-03-10T11:32:09.519196+0000","last_active":"2026-03-10T11:32:13.752686+0000","last_peered":"2026-03-10T11:32:13.752686+0000","last_clean":"2026-03-10T11:32:13.752686+0000","last_became_active":"2026-03-10T11:32:09.519015+0000","last_became_peered":"2026-03-10T11:32:09.519015+0000","last_unstale":"2026-03-10T11:32:13.752686+0000","last_undegraded":"2026-03-10T11:32:13.752686+0000","last_fullsized":"2026-03-10T11:32:13.752686+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:37:26.401819+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,1],"acting":[7,4,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"6.19","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435203+0000","last_change":"2026-03-10T11:32:11.505137+0000","last_active":"2026-03-10T11:32:13.435203+0000","last_peered":"2026-03-10T11:32:13.435203+0000","last_clean":"2026-03-10T11:32:13.435203+0000","last_became_active":"2026-03-10T11:32:11.504903+0000","last_became_peered":"2026-03-10T11:32:11.504903+0000","last_unstale":"2026-03-10T11:32:13.435203+0000","last_undegraded":"2026-03-10T11:32:13.435203+0000","last_fullsized":"2026-03-10T11:32:13.435203+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:14:18.453681+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1,3],"acting":[5,1,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.1e","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059775+0000","last_change":"2026-03-10T11:32:11.576043+0000","last_active":"2026-03-10T11:32:14.059775+0000","last_peered":"2026-03-10T11:32:14.059775+0000","last_clean":"2026-03-10T11:32:14.059775+0000","last_became_active":"2026-03-10T11:32:11.575896+0000","last_became_peered":"2026-03-10T11:32:11.575896+0000","last_unstale":"2026-03-10T11:32:14.059775+0000","last_undegraded":"2026-03-10T11:32:14.059775+0000","last_fullsized":"2026-03-10T11:32:14.059775+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:22:29.629653+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,6,5],"acting":[4,6,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.1b","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539184+0000","last_change":"2026-03-10T11:32:05.494175+0000","last_active":"2026-03-10T11:32:13.539184+0000","last_peered":"2026-03-10T11:32:13.539184+0000","last_clean":"2026-03-10T11:32:13.539184+0000","last_became_active":"2026-03-10T11:32:05.494030+0000","last_became_peered":"2026-03-10T11:32:05.494030+0000","last_unstale":"2026-03-10T11:32:13.539184+0000","last_undegraded":"2026-03-10T11:32:13.539184+0000","last_fullsized":"2026-03-10T11:32:13.539184+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:45:58.156722+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,4,7],"acting":[0,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.1c","version":"55'15","reported_seq":48,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095103+0000","last_change":"2026-03-10T11:32:07.483214+0000","last_active":"2026-03-10T11:32:14.095103+0000","last_peered":"2026-03-10T11:32:14.095103+0000","last_clean":"2026-03-10T11:32:14.095103+0000","last_became_active":"2026-03-10T11:32:07.483035+0000","last_became_peered":"2026-03-10T11:32:07.483035+0000","last_unstale":"2026-03-10T11:32:14.095103+0000","last_undegraded":"2026-03-10T11:32:14.095103+0000","last_fullsized":"2026-03-10T11:32:14.095103+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":15,"log_dups_size":0,"ondisk_log_size":15,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:40:44.403815+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":7,"num_object_clones":0,"num_object_copies":21,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":7,"num_whiteouts":0,"num_read":26,"num_read_kb":17,"num_write":16,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,1,3],"acting":[2,1,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"5.1d","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.756523+0000","last_change":"2026-03-10T11:32:09.519450+0000","last_active":"2026-03-10T11:32:13.756523+0000","last_peered":"2026-03-10T11:32:13.756523+0000","last_clean":"2026-03-10T11:32:13.756523+0000","last_became_active":"2026-03-10T11:32:09.518523+0000","last_became_peered":"2026-03-10T11:32:09.518523+0000","last_unstale":"2026-03-10T11:32:13.756523+0000","last_undegraded":"2026-03-10T11:32:13.756523+0000","last_fullsized":"2026-03-10T11:32:13.756523+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T23:29:05.618724+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,4,0],"acting":[1,4,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"6.1f","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.433035+0000","last_change":"2026-03-10T11:32:11.573792+0000","last_active":"2026-03-10T11:32:13.433035+0000","last_peered":"2026-03-10T11:32:13.433035+0000","last_clean":"2026-03-10T11:32:13.433035+0000","last_became_active":"2026-03-10T11:32:11.573629+0000","last_became_peered":"2026-03-10T11:32:11.573629+0000","last_unstale":"2026-03-10T11:32:13.433035+0000","last_undegraded":"2026-03-10T11:32:13.433035+0000","last_fullsized":"2026-03-10T11:32:13.433035+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:19:57.365774+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,6,5],"acting":[3,6,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.1a","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059058+0000","last_change":"2026-03-10T11:32:05.488341+0000","last_active":"2026-03-10T11:32:14.059058+0000","last_peered":"2026-03-10T11:32:14.059058+0000","last_clean":"2026-03-10T11:32:14.059058+0000","last_became_active":"2026-03-10T11:32:05.488227+0000","last_became_peered":"2026-03-10T11:32:05.488227+0000","last_unstale":"2026-03-10T11:32:14.059058+0000","last_undegraded":"2026-03-10T11:32:14.059058+0000","last_fullsized":"2026-03-10T11:32:14.059058+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:43:35.996147+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,1,2],"acting":[4,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"4.1d","version":"55'12","reported_seq":46,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.433064+0000","last_change":"2026-03-10T11:32:07.492198+0000","last_active":"2026-03-10T11:32:13.433064+0000","last_peered":"2026-03-10T11:32:13.433064+0000","last_clean":"2026-03-10T11:32:13.433064+0000","last_became_active":"2026-03-10T11:32:07.491637+0000","last_became_peered":"2026-03-10T11:32:07.491637+0000","last_unstale":"2026-03-10T11:32:13.433064+0000","last_undegraded":"2026-03-10T11:32:13.433064+0000","last_fullsized":"2026-03-10T11:32:13.433064+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":12,"log_dups_size":0,"ondisk_log_size":12,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:07:54.728750+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":220,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":25,"num_read_kb":16,"num_write":14,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,4],"acting":[3,1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.1c","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059038+0000","last_change":"2026-03-10T11:32:09.515333+0000","last_active":"2026-03-10T11:32:14.059038+0000","last_peered":"2026-03-10T11:32:14.059038+0000","last_clean":"2026-03-10T11:32:14.059038+0000","last_became_active":"2026-03-10T11:32:09.515091+0000","last_became_peered":"2026-03-10T11:32:09.515091+0000","last_unstale":"2026-03-10T11:32:14.059038+0000","last_undegraded":"2026-03-10T11:32:14.059038+0000","last_fullsized":"2026-03-10T11:32:14.059038+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:52:03.100394+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,2],"acting":[4,3,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"6.1c","version":"55'1","reported_seq":16,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753578+0000","last_change":"2026-03-10T11:32:11.514521+0000","last_active":"2026-03-10T11:32:13.753578+0000","last_peered":"2026-03-10T11:32:13.753578+0000","last_clean":"2026-03-10T11:32:13.753578+0000","last_became_active":"2026-03-10T11:32:11.513797+0000","last_became_peered":"2026-03-10T11:32:11.513797+0000","last_unstale":"2026-03-10T11:32:13.753578+0000","last_undegraded":"2026-03-10T11:32:13.753578+0000","last_fullsized":"2026-03-10T11:32:13.753578+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:55:33.816775+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":403,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,5,2],"acting":[7,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"3.19","version":"48'1","reported_seq":28,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.755975+0000","last_change":"2026-03-10T11:32:05.483058+0000","last_active":"2026-03-10T11:32:13.755975+0000","last_peered":"2026-03-10T11:32:13.755975+0000","last_clean":"2026-03-10T11:32:13.755975+0000","last_became_active":"2026-03-10T11:32:05.482922+0000","last_became_peered":"2026-03-10T11:32:05.482922+0000","last_unstale":"2026-03-10T11:32:13.755975+0000","last_undegraded":"2026-03-10T11:32:13.755975+0000","last_fullsized":"2026-03-10T11:32:13.755975+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:08:44.909272+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":46,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":1,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,3,4],"acting":[1,3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"4.1e","version":"55'10","reported_seq":38,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539696+0000","last_change":"2026-03-10T11:32:07.597568+0000","last_active":"2026-03-10T11:32:13.539696+0000","last_peered":"2026-03-10T11:32:13.539696+0000","last_clean":"2026-03-10T11:32:13.539696+0000","last_became_active":"2026-03-10T11:32:07.597489+0000","last_became_peered":"2026-03-10T11:32:07.597489+0000","last_unstale":"2026-03-10T11:32:13.539696+0000","last_undegraded":"2026-03-10T11:32:13.539696+0000","last_fullsized":"2026-03-10T11:32:13.539696+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":10,"log_dups_size":0,"ondisk_log_size":10,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:29:41.419571+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":15,"num_read_kb":10,"num_write":10,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,7,3],"acting":[0,7,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"5.1f","version":"55'8","reported_seq":33,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.430966+0000","last_change":"2026-03-10T11:32:09.517540+0000","last_active":"2026-03-10T11:32:13.430966+0000","last_peered":"2026-03-10T11:32:13.430966+0000","last_clean":"2026-03-10T11:32:13.430966+0000","last_became_active":"2026-03-10T11:32:09.517276+0000","last_became_peered":"2026-03-10T11:32:09.517276+0000","last_unstale":"2026-03-10T11:32:13.430966+0000","last_undegraded":"2026-03-10T11:32:13.430966+0000","last_fullsized":"2026-03-10T11:32:13.430966+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:58:11.682612+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,4,7],"acting":[6,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"4.f","version":"55'15","reported_seq":48,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.756635+0000","last_change":"2026-03-10T11:32:07.487724+0000","last_active":"2026-03-10T11:32:13.756635+0000","last_peered":"2026-03-10T11:32:13.756635+0000","last_clean":"2026-03-10T11:32:13.756635+0000","last_became_active":"2026-03-10T11:32:07.487306+0000","last_became_peered":"2026-03-10T11:32:07.487306+0000","last_unstale":"2026-03-10T11:32:13.756635+0000","last_undegraded":"2026-03-10T11:32:13.756635+0000","last_fullsized":"2026-03-10T11:32:13.756635+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":15,"log_dups_size":0,"ondisk_log_size":15,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:42:46.281406+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":7,"num_object_clones":0,"num_object_copies":21,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":7,"num_whiteouts":0,"num_read":26,"num_read_kb":17,"num_write":16,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,3,4],"acting":[1,3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.8","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432408+0000","last_change":"2026-03-10T11:32:05.492886+0000","last_active":"2026-03-10T11:32:13.432408+0000","last_peered":"2026-03-10T11:32:13.432408+0000","last_clean":"2026-03-10T11:32:13.432408+0000","last_became_active":"2026-03-10T11:32:05.492458+0000","last_became_peered":"2026-03-10T11:32:05.492458+0000","last_unstale":"2026-03-10T11:32:13.432408+0000","last_undegraded":"2026-03-10T11:32:13.432408+0000","last_fullsized":"2026-03-10T11:32:13.432408+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:01:32.990876+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,7],"acting":[3,1,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.e","version":"55'8","reported_seq":29,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058669+0000","last_change":"2026-03-10T11:32:09.515277+0000","last_active":"2026-03-10T11:32:14.058669+0000","last_peered":"2026-03-10T11:32:14.058669+0000","last_clean":"2026-03-10T11:32:14.058669+0000","last_became_active":"2026-03-10T11:32:09.514998+0000","last_became_peered":"2026-03-10T11:32:09.514998+0000","last_unstale":"2026-03-10T11:32:14.058669+0000","last_undegraded":"2026-03-10T11:32:14.058669+0000","last_fullsized":"2026-03-10T11:32:14.058669+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:43:50.105229+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,5,0],"acting":[4,5,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"6.d","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434790+0000","last_change":"2026-03-10T11:32:11.502209+0000","last_active":"2026-03-10T11:32:13.434790+0000","last_peered":"2026-03-10T11:32:13.434790+0000","last_clean":"2026-03-10T11:32:13.434790+0000","last_became_active":"2026-03-10T11:32:11.502106+0000","last_became_peered":"2026-03-10T11:32:11.502106+0000","last_unstale":"2026-03-10T11:32:13.434790+0000","last_undegraded":"2026-03-10T11:32:13.434790+0000","last_fullsized":"2026-03-10T11:32:13.434790+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:42:39.390458+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1,0],"acting":[5,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"4.0","version":"55'18","reported_seq":55,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432508+0000","last_change":"2026-03-10T11:32:07.598762+0000","last_active":"2026-03-10T11:32:13.432508+0000","last_peered":"2026-03-10T11:32:13.432508+0000","last_clean":"2026-03-10T11:32:13.432508+0000","last_became_active":"2026-03-10T11:32:07.598567+0000","last_became_peered":"2026-03-10T11:32:07.598567+0000","last_unstale":"2026-03-10T11:32:13.432508+0000","last_undegraded":"2026-03-10T11:32:13.432508+0000","last_fullsized":"2026-03-10T11:32:13.432508+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":18,"log_dups_size":0,"ondisk_log_size":18,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:45:00.149680+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":220,"num_objects":8,"num_object_clones":0,"num_object_copies":24,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":8,"num_whiteouts":0,"num_read":34,"num_read_kb":22,"num_write":20,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,0],"acting":[3,7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.7","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432476+0000","last_change":"2026-03-10T11:32:05.492977+0000","last_active":"2026-03-10T11:32:13.432476+0000","last_peered":"2026-03-10T11:32:13.432476+0000","last_clean":"2026-03-10T11:32:13.432476+0000","last_became_active":"2026-03-10T11:32:05.492594+0000","last_became_peered":"2026-03-10T11:32:05.492594+0000","last_unstale":"2026-03-10T11:32:13.432476+0000","last_undegraded":"2026-03-10T11:32:13.432476+0000","last_fullsized":"2026-03-10T11:32:13.432476+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:26:21.198593+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,0],"acting":[3,7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.1","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058622+0000","last_change":"2026-03-10T11:32:09.514857+0000","last_active":"2026-03-10T11:32:14.058622+0000","last_peered":"2026-03-10T11:32:14.058622+0000","last_clean":"2026-03-10T11:32:14.058622+0000","last_became_active":"2026-03-10T11:32:09.511003+0000","last_became_peered":"2026-03-10T11:32:09.511003+0000","last_unstale":"2026-03-10T11:32:14.058622+0000","last_undegraded":"2026-03-10T11:32:14.058622+0000","last_fullsized":"2026-03-10T11:32:14.058622+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:51:56.738928+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,7],"acting":[4,3,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"6.2","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058698+0000","last_change":"2026-03-10T11:32:11.510897+0000","last_active":"2026-03-10T11:32:14.058698+0000","last_peered":"2026-03-10T11:32:14.058698+0000","last_clean":"2026-03-10T11:32:14.058698+0000","last_became_active":"2026-03-10T11:32:11.510448+0000","last_became_peered":"2026-03-10T11:32:11.510448+0000","last_unstale":"2026-03-10T11:32:14.058698+0000","last_undegraded":"2026-03-10T11:32:14.058698+0000","last_fullsized":"2026-03-10T11:32:14.058698+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:35:04.941119+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,2],"acting":[4,3,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"4.1","version":"55'14","reported_seq":44,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059961+0000","last_change":"2026-03-10T11:32:07.479250+0000","last_active":"2026-03-10T11:32:14.059961+0000","last_peered":"2026-03-10T11:32:14.059961+0000","last_clean":"2026-03-10T11:32:14.059961+0000","last_became_active":"2026-03-10T11:32:07.478987+0000","last_became_peered":"2026-03-10T11:32:07.478987+0000","last_unstale":"2026-03-10T11:32:14.059961+0000","last_undegraded":"2026-03-10T11:32:14.059961+0000","last_fullsized":"2026-03-10T11:32:14.059961+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":14,"log_dups_size":0,"ondisk_log_size":14,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:14:05.132235+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":7,"num_object_clones":0,"num_object_copies":21,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":7,"num_whiteouts":0,"num_read":21,"num_read_kb":14,"num_write":14,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,5,6],"acting":[4,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.6","version":"48'1","reported_seq":28,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539218+0000","last_change":"2026-03-10T11:32:05.488767+0000","last_active":"2026-03-10T11:32:13.539218+0000","last_peered":"2026-03-10T11:32:13.539218+0000","last_clean":"2026-03-10T11:32:13.539218+0000","last_became_active":"2026-03-10T11:32:05.488670+0000","last_became_peered":"2026-03-10T11:32:05.488670+0000","last_unstale":"2026-03-10T11:32:13.539218+0000","last_undegraded":"2026-03-10T11:32:13.539218+0000","last_fullsized":"2026-03-10T11:32:13.539218+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:38:01.875008+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":46,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":1,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,1,4],"acting":[0,1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"5.0","version":"55'8","reported_seq":29,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431980+0000","last_change":"2026-03-10T11:32:09.495267+0000","last_active":"2026-03-10T11:32:13.431980+0000","last_peered":"2026-03-10T11:32:13.431980+0000","last_clean":"2026-03-10T11:32:13.431980+0000","last_became_active":"2026-03-10T11:32:09.495115+0000","last_became_peered":"2026-03-10T11:32:09.495115+0000","last_unstale":"2026-03-10T11:32:13.431980+0000","last_undegraded":"2026-03-10T11:32:13.431980+0000","last_fullsized":"2026-03-10T11:32:13.431980+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:52:02.237765+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,4],"acting":[3,1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"6.3","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752423+0000","last_change":"2026-03-10T11:32:11.578302+0000","last_active":"2026-03-10T11:32:13.752423+0000","last_peered":"2026-03-10T11:32:13.752423+0000","last_clean":"2026-03-10T11:32:13.752423+0000","last_became_active":"2026-03-10T11:32:11.577452+0000","last_became_peered":"2026-03-10T11:32:11.577452+0000","last_unstale":"2026-03-10T11:32:13.752423+0000","last_undegraded":"2026-03-10T11:32:13.752423+0000","last_fullsized":"2026-03-10T11:32:13.752423+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:24:57.973295+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,6,2],"acting":[7,6,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"4.2","version":"55'10","reported_seq":38,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757315+0000","last_change":"2026-03-10T11:32:07.488057+0000","last_active":"2026-03-10T11:32:13.757315+0000","last_peered":"2026-03-10T11:32:13.757315+0000","last_clean":"2026-03-10T11:32:13.757315+0000","last_became_active":"2026-03-10T11:32:07.487903+0000","last_became_peered":"2026-03-10T11:32:07.487903+0000","last_unstale":"2026-03-10T11:32:13.757315+0000","last_undegraded":"2026-03-10T11:32:13.757315+0000","last_fullsized":"2026-03-10T11:32:13.757315+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":10,"log_dups_size":0,"ondisk_log_size":10,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:22:35.288282+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":15,"num_read_kb":10,"num_write":10,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,4],"acting":[1,5,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.5","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435492+0000","last_change":"2026-03-10T11:32:05.486010+0000","last_active":"2026-03-10T11:32:13.435492+0000","last_peered":"2026-03-10T11:32:13.435492+0000","last_clean":"2026-03-10T11:32:13.435492+0000","last_became_active":"2026-03-10T11:32:05.485734+0000","last_became_peered":"2026-03-10T11:32:05.485734+0000","last_unstale":"2026-03-10T11:32:13.435492+0000","last_undegraded":"2026-03-10T11:32:13.435492+0000","last_fullsized":"2026-03-10T11:32:13.435492+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:58:01.385548+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,3,2],"acting":[5,3,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"5.3","version":"55'8","reported_seq":30,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539874+0000","last_change":"2026-03-10T11:32:09.505558+0000","last_active":"2026-03-10T11:32:13.539874+0000","last_peered":"2026-03-10T11:32:13.539874+0000","last_clean":"2026-03-10T11:32:13.539874+0000","last_became_active":"2026-03-10T11:32:09.505467+0000","last_became_peered":"2026-03-10T11:32:09.505467+0000","last_unstale":"2026-03-10T11:32:13.539874+0000","last_undegraded":"2026-03-10T11:32:13.539874+0000","last_fullsized":"2026-03-10T11:32:13.539874+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:27:40.165203+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,6,5],"acting":[0,6,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"6.0","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539722+0000","last_change":"2026-03-10T11:32:11.524869+0000","last_active":"2026-03-10T11:32:13.539722+0000","last_peered":"2026-03-10T11:32:13.539722+0000","last_clean":"2026-03-10T11:32:13.539722+0000","last_became_active":"2026-03-10T11:32:11.524764+0000","last_became_peered":"2026-03-10T11:32:11.524764+0000","last_unstale":"2026-03-10T11:32:13.539722+0000","last_undegraded":"2026-03-10T11:32:13.539722+0000","last_fullsized":"2026-03-10T11:32:13.539722+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:50:58.880334+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,3,2],"acting":[0,3,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.3","version":"55'19","reported_seq":59,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539585+0000","last_change":"2026-03-10T11:32:07.597361+0000","last_active":"2026-03-10T11:32:13.539585+0000","last_peered":"2026-03-10T11:32:13.539585+0000","last_clean":"2026-03-10T11:32:13.539585+0000","last_became_active":"2026-03-10T11:32:07.597199+0000","last_became_peered":"2026-03-10T11:32:07.597199+0000","last_unstale":"2026-03-10T11:32:13.539585+0000","last_undegraded":"2026-03-10T11:32:13.539585+0000","last_fullsized":"2026-03-10T11:32:13.539585+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":19,"log_dups_size":0,"ondisk_log_size":19,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:31:09.996144+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":330,"num_objects":8,"num_object_clones":0,"num_object_copies":24,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":8,"num_whiteouts":0,"num_read":39,"num_read_kb":25,"num_write":22,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,5,7],"acting":[0,5,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"3.4","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.755928+0000","last_change":"2026-03-10T11:32:05.470777+0000","last_active":"2026-03-10T11:32:13.755928+0000","last_peered":"2026-03-10T11:32:13.755928+0000","last_clean":"2026-03-10T11:32:13.755928+0000","last_became_active":"2026-03-10T11:32:05.469871+0000","last_became_peered":"2026-03-10T11:32:05.469871+0000","last_unstale":"2026-03-10T11:32:13.755928+0000","last_undegraded":"2026-03-10T11:32:13.755928+0000","last_fullsized":"2026-03-10T11:32:13.755928+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:40:37.152961+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,5],"acting":[1,2,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"5.2","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431003+0000","last_change":"2026-03-10T11:32:09.517469+0000","last_active":"2026-03-10T11:32:13.431003+0000","last_peered":"2026-03-10T11:32:13.431003+0000","last_clean":"2026-03-10T11:32:13.431003+0000","last_became_active":"2026-03-10T11:32:09.517068+0000","last_became_peered":"2026-03-10T11:32:09.517068+0000","last_unstale":"2026-03-10T11:32:13.431003+0000","last_undegraded":"2026-03-10T11:32:13.431003+0000","last_fullsized":"2026-03-10T11:32:13.431003+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:59:19.119041+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,0,5],"acting":[6,0,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"6.1","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757256+0000","last_change":"2026-03-10T11:32:11.575149+0000","last_active":"2026-03-10T11:32:13.757256+0000","last_peered":"2026-03-10T11:32:13.757256+0000","last_clean":"2026-03-10T11:32:13.757256+0000","last_became_active":"2026-03-10T11:32:11.574450+0000","last_became_peered":"2026-03-10T11:32:11.574450+0000","last_unstale":"2026-03-10T11:32:13.757256+0000","last_undegraded":"2026-03-10T11:32:13.757256+0000","last_fullsized":"2026-03-10T11:32:13.757256+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:44:00.507675+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,6,2],"acting":[1,6,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"4.4","version":"55'28","reported_seq":73,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757031+0000","last_change":"2026-03-10T11:32:07.486298+0000","last_active":"2026-03-10T11:32:13.757031+0000","last_peered":"2026-03-10T11:32:13.757031+0000","last_clean":"2026-03-10T11:32:13.757031+0000","last_became_active":"2026-03-10T11:32:07.486212+0000","last_became_peered":"2026-03-10T11:32:07.486212+0000","last_unstale":"2026-03-10T11:32:13.757031+0000","last_undegraded":"2026-03-10T11:32:13.757031+0000","last_fullsized":"2026-03-10T11:32:13.757031+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":28,"log_dups_size":0,"ondisk_log_size":28,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:37:25.312706+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":358,"num_objects":10,"num_object_clones":0,"num_object_copies":30,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":10,"num_whiteouts":0,"num_read":48,"num_read_kb":33,"num_write":26,"num_write_kb":4,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,3],"acting":[1,2,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.3","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058290+0000","last_change":"2026-03-10T11:32:05.482284+0000","last_active":"2026-03-10T11:32:14.058290+0000","last_peered":"2026-03-10T11:32:14.058290+0000","last_clean":"2026-03-10T11:32:14.058290+0000","last_became_active":"2026-03-10T11:32:05.482174+0000","last_became_peered":"2026-03-10T11:32:05.482174+0000","last_unstale":"2026-03-10T11:32:14.058290+0000","last_undegraded":"2026-03-10T11:32:14.058290+0000","last_fullsized":"2026-03-10T11:32:14.058290+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:28:56.826046+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,0,6],"acting":[4,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"2.2","version":"50'2","reported_seq":34,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435583+0000","last_change":"2026-03-10T11:32:07.468273+0000","last_active":"2026-03-10T11:32:13.435583+0000","last_peered":"2026-03-10T11:32:13.435583+0000","last_clean":"2026-03-10T11:32:13.435583+0000","last_became_active":"2026-03-10T11:32:05.466762+0000","last_became_peered":"2026-03-10T11:32:05.466762+0000","last_unstale":"2026-03-10T11:32:13.435583+0000","last_undegraded":"2026-03-10T11:32:13.435583+0000","last_fullsized":"2026-03-10T11:32:13.435583+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:51:31.016672+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00040810500000000002,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1,6],"acting":[5,1,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"5.5","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539487+0000","last_change":"2026-03-10T11:32:09.515802+0000","last_active":"2026-03-10T11:32:13.539487+0000","last_peered":"2026-03-10T11:32:13.539487+0000","last_clean":"2026-03-10T11:32:13.539487+0000","last_became_active":"2026-03-10T11:32:09.515691+0000","last_became_peered":"2026-03-10T11:32:09.515691+0000","last_unstale":"2026-03-10T11:32:13.539487+0000","last_undegraded":"2026-03-10T11:32:13.539487+0000","last_fullsized":"2026-03-10T11:32:13.539487+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:34:27.429654+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,1,4],"acting":[0,1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"6.6","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432064+0000","last_change":"2026-03-10T11:32:11.509752+0000","last_active":"2026-03-10T11:32:13.432064+0000","last_peered":"2026-03-10T11:32:13.432064+0000","last_clean":"2026-03-10T11:32:13.432064+0000","last_became_active":"2026-03-10T11:32:11.509288+0000","last_became_peered":"2026-03-10T11:32:11.509288+0000","last_unstale":"2026-03-10T11:32:13.432064+0000","last_undegraded":"2026-03-10T11:32:13.432064+0000","last_fullsized":"2026-03-10T11:32:13.432064+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:25:33.241815+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,4,7],"acting":[3,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.7","version":"55'13","reported_seq":50,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757287+0000","last_change":"2026-03-10T11:32:07.485025+0000","last_active":"2026-03-10T11:32:13.757287+0000","last_peered":"2026-03-10T11:32:13.757287+0000","last_clean":"2026-03-10T11:32:13.757287+0000","last_became_active":"2026-03-10T11:32:07.484944+0000","last_became_peered":"2026-03-10T11:32:07.484944+0000","last_unstale":"2026-03-10T11:32:13.757287+0000","last_undegraded":"2026-03-10T11:32:13.757287+0000","last_fullsized":"2026-03-10T11:32:13.757287+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":13,"log_dups_size":0,"ondisk_log_size":13,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:49:17.507119+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":330,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":30,"num_read_kb":19,"num_write":16,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,0],"acting":[1,5,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.0","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.755868+0000","last_change":"2026-03-10T11:32:05.469815+0000","last_active":"2026-03-10T11:32:13.755868+0000","last_peered":"2026-03-10T11:32:13.755868+0000","last_clean":"2026-03-10T11:32:13.755868+0000","last_became_active":"2026-03-10T11:32:05.469717+0000","last_became_peered":"2026-03-10T11:32:05.469717+0000","last_unstale":"2026-03-10T11:32:13.755868+0000","last_undegraded":"2026-03-10T11:32:13.755868+0000","last_fullsized":"2026-03-10T11:32:13.755868+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:23:39.249935+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,6],"acting":[1,2,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"2.1","version":"48'1","reported_seq":33,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.094602+0000","last_change":"2026-03-10T11:32:07.469162+0000","last_active":"2026-03-10T11:32:14.094602+0000","last_peered":"2026-03-10T11:32:14.094602+0000","last_clean":"2026-03-10T11:32:14.094602+0000","last_became_active":"2026-03-10T11:32:05.481644+0000","last_became_peered":"2026-03-10T11:32:05.481644+0000","last_unstale":"2026-03-10T11:32:14.094602+0000","last_undegraded":"2026-03-10T11:32:14.094602+0000","last_fullsized":"2026-03-10T11:32:14.094602+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:54:36.058555+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00034258999999999998,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,3,0],"acting":[2,3,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"5.6","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095075+0000","last_change":"2026-03-10T11:32:09.504060+0000","last_active":"2026-03-10T11:32:14.095075+0000","last_peered":"2026-03-10T11:32:14.095075+0000","last_clean":"2026-03-10T11:32:14.095075+0000","last_became_active":"2026-03-10T11:32:09.503889+0000","last_became_peered":"2026-03-10T11:32:09.503889+0000","last_unstale":"2026-03-10T11:32:14.095075+0000","last_undegraded":"2026-03-10T11:32:14.095075+0000","last_fullsized":"2026-03-10T11:32:14.095075+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:29:57.323993+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,5,7],"acting":[2,5,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"6.5","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752569+0000","last_change":"2026-03-10T11:32:11.577427+0000","last_active":"2026-03-10T11:32:13.752569+0000","last_peered":"2026-03-10T11:32:13.752569+0000","last_clean":"2026-03-10T11:32:13.752569+0000","last_became_active":"2026-03-10T11:32:11.576376+0000","last_became_peered":"2026-03-10T11:32:11.576376+0000","last_unstale":"2026-03-10T11:32:13.752569+0000","last_undegraded":"2026-03-10T11:32:13.752569+0000","last_fullsized":"2026-03-10T11:32:13.752569+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:12:59.201772+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,6,3],"acting":[7,6,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"4.6","version":"55'12","reported_seq":41,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539292+0000","last_change":"2026-03-10T11:32:07.491133+0000","last_active":"2026-03-10T11:32:13.539292+0000","last_peered":"2026-03-10T11:32:13.539292+0000","last_clean":"2026-03-10T11:32:13.539292+0000","last_became_active":"2026-03-10T11:32:07.490960+0000","last_became_peered":"2026-03-10T11:32:07.490960+0000","last_unstale":"2026-03-10T11:32:13.539292+0000","last_undegraded":"2026-03-10T11:32:13.539292+0000","last_fullsized":"2026-03-10T11:32:13.539292+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":12,"log_dups_size":0,"ondisk_log_size":12,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:52:08.881689+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":6,"num_object_clones":0,"num_object_copies":18,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":6,"num_whiteouts":0,"num_read":18,"num_read_kb":12,"num_write":12,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,1,2],"acting":[0,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"3.1","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.538897+0000","last_change":"2026-03-10T11:32:05.484386+0000","last_active":"2026-03-10T11:32:13.538897+0000","last_peered":"2026-03-10T11:32:13.538897+0000","last_clean":"2026-03-10T11:32:13.538897+0000","last_became_active":"2026-03-10T11:32:05.484276+0000","last_became_peered":"2026-03-10T11:32:05.484276+0000","last_unstale":"2026-03-10T11:32:13.538897+0000","last_undegraded":"2026-03-10T11:32:13.538897+0000","last_fullsized":"2026-03-10T11:32:13.538897+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T23:27:09.503732+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,4,3],"acting":[0,4,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"2.0","version":"56'5","reported_seq":43,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:15.198196+0000","last_change":"2026-03-10T11:32:07.595732+0000","last_active":"2026-03-10T11:32:15.198196+0000","last_peered":"2026-03-10T11:32:15.198196+0000","last_clean":"2026-03-10T11:32:15.198196+0000","last_became_active":"2026-03-10T11:32:05.484114+0000","last_became_peered":"2026-03-10T11:32:05.484114+0000","last_unstale":"2026-03-10T11:32:15.198196+0000","last_undegraded":"2026-03-10T11:32:15.198196+0000","last_fullsized":"2026-03-10T11:32:15.198196+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":5,"log_dups_size":0,"ondisk_log_size":5,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:21:40.405562+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00025977500000000001,"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":8,"num_read_kb":3,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,1,0],"acting":[7,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.7","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435296+0000","last_change":"2026-03-10T11:32:09.500766+0000","last_active":"2026-03-10T11:32:13.435296+0000","last_peered":"2026-03-10T11:32:13.435296+0000","last_clean":"2026-03-10T11:32:13.435296+0000","last_became_active":"2026-03-10T11:32:09.500135+0000","last_became_peered":"2026-03-10T11:32:09.500135+0000","last_unstale":"2026-03-10T11:32:13.435296+0000","last_undegraded":"2026-03-10T11:32:13.435296+0000","last_fullsized":"2026-03-10T11:32:13.435296+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:02:55.101840+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1,0],"acting":[5,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.4","version":"55'1","reported_seq":16,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.756803+0000","last_change":"2026-03-10T11:32:11.515730+0000","last_active":"2026-03-10T11:32:13.756803+0000","last_peered":"2026-03-10T11:32:13.756803+0000","last_clean":"2026-03-10T11:32:13.756803+0000","last_became_active":"2026-03-10T11:32:11.511711+0000","last_became_peered":"2026-03-10T11:32:11.511711+0000","last_unstale":"2026-03-10T11:32:13.756803+0000","last_undegraded":"2026-03-10T11:32:13.756803+0000","last_fullsized":"2026-03-10T11:32:13.756803+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:39:45.795505+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":13,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":1,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,3],"acting":[1,5,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"4.5","version":"55'16","reported_seq":48,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431268+0000","last_change":"2026-03-10T11:32:07.597624+0000","last_active":"2026-03-10T11:32:13.431268+0000","last_peered":"2026-03-10T11:32:13.431268+0000","last_clean":"2026-03-10T11:32:13.431268+0000","last_became_active":"2026-03-10T11:32:07.597049+0000","last_became_peered":"2026-03-10T11:32:07.597049+0000","last_unstale":"2026-03-10T11:32:13.431268+0000","last_undegraded":"2026-03-10T11:32:13.431268+0000","last_fullsized":"2026-03-10T11:32:13.431268+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":16,"log_dups_size":0,"ondisk_log_size":16,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:17:30.419013+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":154,"num_objects":8,"num_object_clones":0,"num_object_copies":24,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":8,"num_whiteouts":0,"num_read":25,"num_read_kb":15,"num_write":13,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,4,7],"acting":[6,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"3.2","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432591+0000","last_change":"2026-03-10T11:32:05.483910+0000","last_active":"2026-03-10T11:32:13.432591+0000","last_peered":"2026-03-10T11:32:13.432591+0000","last_clean":"2026-03-10T11:32:13.432591+0000","last_became_active":"2026-03-10T11:32:05.480854+0000","last_became_peered":"2026-03-10T11:32:05.480854+0000","last_unstale":"2026-03-10T11:32:13.432591+0000","last_undegraded":"2026-03-10T11:32:13.432591+0000","last_fullsized":"2026-03-10T11:32:13.432591+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:44:44.275870+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,6],"acting":[3,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"1.0","version":"18'32","reported_seq":37,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753175+0000","last_change":"2026-03-10T11:32:03.457270+0000","last_active":"2026-03-10T11:32:13.753175+0000","last_peered":"2026-03-10T11:32:13.753175+0000","last_clean":"2026-03-10T11:32:13.753175+0000","last_became_active":"2026-03-10T11:32:03.451510+0000","last_became_peered":"2026-03-10T11:32:03.451510+0000","last_unstale":"2026-03-10T11:32:13.753175+0000","last_undegraded":"2026-03-10T11:32:13.753175+0000","last_fullsized":"2026-03-10T11:32:13.753175+0000","mapping_epoch":45,"log_start":"0'0","ondisk_log_start":"0'0","created":17,"last_epoch_clean":46,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:31:09.189169+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:31:09.189169+0000","last_clean_scrub_stamp":"2026-03-10T11:31:09.189169+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:34:21.433105+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":459280,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0,6],"acting":[7,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.4","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753235+0000","last_change":"2026-03-10T11:32:09.525158+0000","last_active":"2026-03-10T11:32:13.753235+0000","last_peered":"2026-03-10T11:32:13.753235+0000","last_clean":"2026-03-10T11:32:13.753235+0000","last_became_active":"2026-03-10T11:32:09.525051+0000","last_became_peered":"2026-03-10T11:32:09.525051+0000","last_unstale":"2026-03-10T11:32:13.753235+0000","last_undegraded":"2026-03-10T11:32:13.753235+0000","last_fullsized":"2026-03-10T11:32:13.753235+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:51:00.114489+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,2,5],"acting":[7,2,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"6.7","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434911+0000","last_change":"2026-03-10T11:32:11.505220+0000","last_active":"2026-03-10T11:32:13.434911+0000","last_peered":"2026-03-10T11:32:13.434911+0000","last_clean":"2026-03-10T11:32:13.434911+0000","last_became_active":"2026-03-10T11:32:11.505041+0000","last_became_peered":"2026-03-10T11:32:11.505041+0000","last_unstale":"2026-03-10T11:32:13.434911+0000","last_undegraded":"2026-03-10T11:32:13.434911+0000","last_fullsized":"2026-03-10T11:32:13.434911+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:20:43.846661+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,3,4],"acting":[5,3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"4.e","version":"55'11","reported_seq":42,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059204+0000","last_change":"2026-03-10T11:32:07.482751+0000","last_active":"2026-03-10T11:32:14.059204+0000","last_peered":"2026-03-10T11:32:14.059204+0000","last_clean":"2026-03-10T11:32:14.059204+0000","last_became_active":"2026-03-10T11:32:07.482298+0000","last_became_peered":"2026-03-10T11:32:07.482298+0000","last_unstale":"2026-03-10T11:32:14.059204+0000","last_undegraded":"2026-03-10T11:32:14.059204+0000","last_fullsized":"2026-03-10T11:32:14.059204+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":11,"log_dups_size":0,"ondisk_log_size":11,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:29:34.576488+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":20,"num_read_kb":13,"num_write":12,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,6,1],"acting":[4,6,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.9","version":"48'1","reported_seq":33,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059171+0000","last_change":"2026-03-10T11:32:05.492929+0000","last_active":"2026-03-10T11:32:14.059171+0000","last_peered":"2026-03-10T11:32:14.059171+0000","last_clean":"2026-03-10T11:32:14.059171+0000","last_became_active":"2026-03-10T11:32:05.492820+0000","last_became_peered":"2026-03-10T11:32:05.492820+0000","last_unstale":"2026-03-10T11:32:14.059171+0000","last_undegraded":"2026-03-10T11:32:14.059171+0000","last_fullsized":"2026-03-10T11:32:14.059171+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:00:17.535577+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":993,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":7,"num_read_kb":7,"num_write":2,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,2,7],"acting":[4,2,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"5.f","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435749+0000","last_change":"2026-03-10T11:32:09.513230+0000","last_active":"2026-03-10T11:32:13.435749+0000","last_peered":"2026-03-10T11:32:13.435749+0000","last_clean":"2026-03-10T11:32:13.435749+0000","last_became_active":"2026-03-10T11:32:09.513045+0000","last_became_peered":"2026-03-10T11:32:09.513045+0000","last_unstale":"2026-03-10T11:32:13.435749+0000","last_undegraded":"2026-03-10T11:32:13.435749+0000","last_fullsized":"2026-03-10T11:32:13.435749+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:48:47.617831+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,4,6],"acting":[5,4,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.c","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431921+0000","last_change":"2026-03-10T11:32:11.573142+0000","last_active":"2026-03-10T11:32:13.431921+0000","last_peered":"2026-03-10T11:32:13.431921+0000","last_clean":"2026-03-10T11:32:13.431921+0000","last_became_active":"2026-03-10T11:32:11.573026+0000","last_became_peered":"2026-03-10T11:32:11.573026+0000","last_unstale":"2026-03-10T11:32:13.431921+0000","last_undegraded":"2026-03-10T11:32:13.431921+0000","last_fullsized":"2026-03-10T11:32:13.431921+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:18:04.964512+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,6,5],"acting":[3,6,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.d","version":"55'17","reported_seq":51,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059505+0000","last_change":"2026-03-10T11:32:07.482964+0000","last_active":"2026-03-10T11:32:14.059505+0000","last_peered":"2026-03-10T11:32:14.059505+0000","last_clean":"2026-03-10T11:32:14.059505+0000","last_became_active":"2026-03-10T11:32:07.482540+0000","last_became_peered":"2026-03-10T11:32:07.482540+0000","last_unstale":"2026-03-10T11:32:14.059505+0000","last_undegraded":"2026-03-10T11:32:14.059505+0000","last_fullsized":"2026-03-10T11:32:14.059505+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":17,"log_dups_size":0,"ondisk_log_size":17,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:51:57.573737+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":8,"num_object_clones":0,"num_object_copies":24,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":8,"num_whiteouts":0,"num_read":29,"num_read_kb":19,"num_write":18,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,2,1],"acting":[4,2,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.a","version":"48'1","reported_seq":33,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431472+0000","last_change":"2026-03-10T11:32:05.482905+0000","last_active":"2026-03-10T11:32:13.431472+0000","last_peered":"2026-03-10T11:32:13.431472+0000","last_clean":"2026-03-10T11:32:13.431472+0000","last_became_active":"2026-03-10T11:32:05.482830+0000","last_became_peered":"2026-03-10T11:32:05.482830+0000","last_unstale":"2026-03-10T11:32:13.431472+0000","last_undegraded":"2026-03-10T11:32:13.431472+0000","last_fullsized":"2026-03-10T11:32:13.431472+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:41:28.409216+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":436,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":7,"num_read_kb":7,"num_write":2,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,4,1],"acting":[6,4,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"5.c","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.756856+0000","last_change":"2026-03-10T11:32:09.518756+0000","last_active":"2026-03-10T11:32:13.756856+0000","last_peered":"2026-03-10T11:32:13.756856+0000","last_clean":"2026-03-10T11:32:13.756856+0000","last_became_active":"2026-03-10T11:32:09.518665+0000","last_became_peered":"2026-03-10T11:32:09.518665+0000","last_unstale":"2026-03-10T11:32:13.756856+0000","last_undegraded":"2026-03-10T11:32:13.756856+0000","last_fullsized":"2026-03-10T11:32:13.756856+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:18:11.441776+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,4,0],"acting":[1,4,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"6.f","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095132+0000","last_change":"2026-03-10T11:32:11.527689+0000","last_active":"2026-03-10T11:32:14.095132+0000","last_peered":"2026-03-10T11:32:14.095132+0000","last_clean":"2026-03-10T11:32:14.095132+0000","last_became_active":"2026-03-10T11:32:11.527559+0000","last_became_peered":"2026-03-10T11:32:11.527559+0000","last_unstale":"2026-03-10T11:32:14.095132+0000","last_undegraded":"2026-03-10T11:32:14.095132+0000","last_fullsized":"2026-03-10T11:32:14.095132+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:29:29.018133+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,3,4],"acting":[2,3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"4.c","version":"55'10","reported_seq":38,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059742+0000","last_change":"2026-03-10T11:32:07.481053+0000","last_active":"2026-03-10T11:32:14.059742+0000","last_peered":"2026-03-10T11:32:14.059742+0000","last_clean":"2026-03-10T11:32:14.059742+0000","last_became_active":"2026-03-10T11:32:07.480920+0000","last_became_peered":"2026-03-10T11:32:07.480920+0000","last_unstale":"2026-03-10T11:32:14.059742+0000","last_undegraded":"2026-03-10T11:32:14.059742+0000","last_fullsized":"2026-03-10T11:32:14.059742+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":10,"log_dups_size":0,"ondisk_log_size":10,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:44:15.363985+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":15,"num_read_kb":10,"num_write":10,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,6],"acting":[4,3,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.b","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432351+0000","last_change":"2026-03-10T11:32:05.484342+0000","last_active":"2026-03-10T11:32:13.432351+0000","last_peered":"2026-03-10T11:32:13.432351+0000","last_clean":"2026-03-10T11:32:13.432351+0000","last_became_active":"2026-03-10T11:32:05.484242+0000","last_became_peered":"2026-03-10T11:32:05.484242+0000","last_unstale":"2026-03-10T11:32:13.432351+0000","last_undegraded":"2026-03-10T11:32:13.432351+0000","last_fullsized":"2026-03-10T11:32:13.432351+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:32:44.695262+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0,4],"acting":[3,0,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.d","version":"55'8","reported_seq":29,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095343+0000","last_change":"2026-03-10T11:32:09.504159+0000","last_active":"2026-03-10T11:32:14.095343+0000","last_peered":"2026-03-10T11:32:14.095343+0000","last_clean":"2026-03-10T11:32:14.095343+0000","last_became_active":"2026-03-10T11:32:09.503998+0000","last_became_peered":"2026-03-10T11:32:09.503998+0000","last_unstale":"2026-03-10T11:32:14.095343+0000","last_undegraded":"2026-03-10T11:32:14.095343+0000","last_fullsized":"2026-03-10T11:32:14.095343+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:09:16.830610+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,7,5],"acting":[2,7,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"6.e","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059642+0000","last_change":"2026-03-10T11:32:11.512156+0000","last_active":"2026-03-10T11:32:14.059642+0000","last_peered":"2026-03-10T11:32:14.059642+0000","last_clean":"2026-03-10T11:32:14.059642+0000","last_became_active":"2026-03-10T11:32:11.512023+0000","last_became_peered":"2026-03-10T11:32:11.512023+0000","last_unstale":"2026-03-10T11:32:14.059642+0000","last_undegraded":"2026-03-10T11:32:14.059642+0000","last_fullsized":"2026-03-10T11:32:14.059642+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:31:40.808175+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,1,2],"acting":[4,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"4.b","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539640+0000","last_change":"2026-03-10T11:32:07.490958+0000","last_active":"2026-03-10T11:32:13.539640+0000","last_peered":"2026-03-10T11:32:13.539640+0000","last_clean":"2026-03-10T11:32:13.539640+0000","last_became_active":"2026-03-10T11:32:07.490846+0000","last_became_peered":"2026-03-10T11:32:07.490846+0000","last_unstale":"2026-03-10T11:32:13.539640+0000","last_undegraded":"2026-03-10T11:32:13.539640+0000","last_fullsized":"2026-03-10T11:32:13.539640+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:02:19.295560+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,1,4],"acting":[0,1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"3.c","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435404+0000","last_change":"2026-03-10T11:32:05.486072+0000","last_active":"2026-03-10T11:32:13.435404+0000","last_peered":"2026-03-10T11:32:13.435404+0000","last_clean":"2026-03-10T11:32:13.435404+0000","last_became_active":"2026-03-10T11:32:05.485842+0000","last_became_peered":"2026-03-10T11:32:05.485842+0000","last_unstale":"2026-03-10T11:32:13.435404+0000","last_undegraded":"2026-03-10T11:32:13.435404+0000","last_fullsized":"2026-03-10T11:32:13.435404+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:34:56.952419+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,3,6],"acting":[5,3,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"5.a","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095262+0000","last_change":"2026-03-10T11:32:09.492441+0000","last_active":"2026-03-10T11:32:14.095262+0000","last_peered":"2026-03-10T11:32:14.095262+0000","last_clean":"2026-03-10T11:32:14.095262+0000","last_became_active":"2026-03-10T11:32:09.492151+0000","last_became_peered":"2026-03-10T11:32:09.492151+0000","last_unstale":"2026-03-10T11:32:14.095262+0000","last_undegraded":"2026-03-10T11:32:14.095262+0000","last_fullsized":"2026-03-10T11:32:14.095262+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:16:19.045768+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,4,3],"acting":[2,4,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"6.9","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539613+0000","last_change":"2026-03-10T11:32:11.502345+0000","last_active":"2026-03-10T11:32:13.539613+0000","last_peered":"2026-03-10T11:32:13.539613+0000","last_clean":"2026-03-10T11:32:13.539613+0000","last_became_active":"2026-03-10T11:32:11.502116+0000","last_became_peered":"2026-03-10T11:32:11.502116+0000","last_unstale":"2026-03-10T11:32:13.539613+0000","last_undegraded":"2026-03-10T11:32:13.539613+0000","last_fullsized":"2026-03-10T11:32:13.539613+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:16:04.120385+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,7,2],"acting":[0,7,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.a","version":"55'19","reported_seq":54,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431517+0000","last_change":"2026-03-10T11:32:07.597136+0000","last_active":"2026-03-10T11:32:13.431517+0000","last_peered":"2026-03-10T11:32:13.431517+0000","last_clean":"2026-03-10T11:32:13.431517+0000","last_became_active":"2026-03-10T11:32:07.596902+0000","last_became_peered":"2026-03-10T11:32:07.596902+0000","last_unstale":"2026-03-10T11:32:13.431517+0000","last_undegraded":"2026-03-10T11:32:13.431517+0000","last_fullsized":"2026-03-10T11:32:13.431517+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":19,"log_dups_size":0,"ondisk_log_size":19,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:06:20.801200+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":9,"num_object_clones":0,"num_object_copies":27,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":9,"num_whiteouts":0,"num_read":32,"num_read_kb":21,"num_write":20,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,1,7],"acting":[6,1,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"3.d","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752240+0000","last_change":"2026-03-10T11:32:05.484420+0000","last_active":"2026-03-10T11:32:13.752240+0000","last_peered":"2026-03-10T11:32:13.752240+0000","last_clean":"2026-03-10T11:32:13.752240+0000","last_became_active":"2026-03-10T11:32:05.484301+0000","last_became_peered":"2026-03-10T11:32:05.484301+0000","last_unstale":"2026-03-10T11:32:13.752240+0000","last_undegraded":"2026-03-10T11:32:13.752240+0000","last_fullsized":"2026-03-10T11:32:13.752240+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:56:27.516701+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,5,6],"acting":[7,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.b","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095302+0000","last_change":"2026-03-10T11:32:09.497693+0000","last_active":"2026-03-10T11:32:14.095302+0000","last_peered":"2026-03-10T11:32:14.095302+0000","last_clean":"2026-03-10T11:32:14.095302+0000","last_became_active":"2026-03-10T11:32:09.497549+0000","last_became_peered":"2026-03-10T11:32:09.497549+0000","last_unstale":"2026-03-10T11:32:14.095302+0000","last_undegraded":"2026-03-10T11:32:14.095302+0000","last_fullsized":"2026-03-10T11:32:14.095302+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T23:07:22.489823+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,0,5],"acting":[2,0,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"6.8","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753123+0000","last_change":"2026-03-10T11:32:11.512386+0000","last_active":"2026-03-10T11:32:13.753123+0000","last_peered":"2026-03-10T11:32:13.753123+0000","last_clean":"2026-03-10T11:32:13.753123+0000","last_became_active":"2026-03-10T11:32:11.511538+0000","last_became_peered":"2026-03-10T11:32:11.511538+0000","last_unstale":"2026-03-10T11:32:13.753123+0000","last_undegraded":"2026-03-10T11:32:13.753123+0000","last_fullsized":"2026-03-10T11:32:13.753123+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:48:17.780752+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,2,3],"acting":[7,2,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"4.9","version":"55'12","reported_seq":46,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059256+0000","last_change":"2026-03-10T11:32:07.483133+0000","last_active":"2026-03-10T11:32:14.059256+0000","last_peered":"2026-03-10T11:32:14.059256+0000","last_clean":"2026-03-10T11:32:14.059256+0000","last_became_active":"2026-03-10T11:32:07.482407+0000","last_became_peered":"2026-03-10T11:32:07.482407+0000","last_unstale":"2026-03-10T11:32:14.059256+0000","last_undegraded":"2026-03-10T11:32:14.059256+0000","last_fullsized":"2026-03-10T11:32:14.059256+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":12,"log_dups_size":0,"ondisk_log_size":12,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:07:32.679946+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":220,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":25,"num_read_kb":16,"num_write":14,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,1,3],"acting":[4,1,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.e","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753020+0000","last_change":"2026-03-10T11:32:05.479546+0000","last_active":"2026-03-10T11:32:13.753020+0000","last_peered":"2026-03-10T11:32:13.753020+0000","last_clean":"2026-03-10T11:32:13.753020+0000","last_became_active":"2026-03-10T11:32:05.479050+0000","last_became_peered":"2026-03-10T11:32:05.479050+0000","last_unstale":"2026-03-10T11:32:13.753020+0000","last_undegraded":"2026-03-10T11:32:13.753020+0000","last_fullsized":"2026-03-10T11:32:13.753020+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:33:49.505723+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,1],"acting":[7,4,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.8","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095213+0000","last_change":"2026-03-10T11:32:09.498834+0000","last_active":"2026-03-10T11:32:14.095213+0000","last_peered":"2026-03-10T11:32:14.095213+0000","last_clean":"2026-03-10T11:32:14.095213+0000","last_became_active":"2026-03-10T11:32:09.498594+0000","last_became_peered":"2026-03-10T11:32:09.498594+0000","last_unstale":"2026-03-10T11:32:14.095213+0000","last_undegraded":"2026-03-10T11:32:14.095213+0000","last_fullsized":"2026-03-10T11:32:14.095213+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:10:52.238205+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,0,1],"acting":[2,0,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"6.b","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.433102+0000","last_change":"2026-03-10T11:32:11.510427+0000","last_active":"2026-03-10T11:32:13.433102+0000","last_peered":"2026-03-10T11:32:13.433102+0000","last_clean":"2026-03-10T11:32:13.433102+0000","last_became_active":"2026-03-10T11:32:11.510285+0000","last_became_peered":"2026-03-10T11:32:11.510285+0000","last_unstale":"2026-03-10T11:32:13.433102+0000","last_undegraded":"2026-03-10T11:32:13.433102+0000","last_fullsized":"2026-03-10T11:32:13.433102+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:33:36.994107+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,1],"acting":[3,7,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.8","version":"55'15","reported_seq":48,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434844+0000","last_change":"2026-03-10T11:32:07.597775+0000","last_active":"2026-03-10T11:32:13.434844+0000","last_peered":"2026-03-10T11:32:13.434844+0000","last_clean":"2026-03-10T11:32:13.434844+0000","last_became_active":"2026-03-10T11:32:07.597555+0000","last_became_peered":"2026-03-10T11:32:07.597555+0000","last_unstale":"2026-03-10T11:32:13.434844+0000","last_undegraded":"2026-03-10T11:32:13.434844+0000","last_fullsized":"2026-03-10T11:32:13.434844+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":15,"log_dups_size":0,"ondisk_log_size":15,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:55:39.053348+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":7,"num_object_clones":0,"num_object_copies":21,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":7,"num_whiteouts":0,"num_read":26,"num_read_kb":17,"num_write":16,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,7,6],"acting":[5,7,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"3.f","version":"48'2","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752919+0000","last_change":"2026-03-10T11:32:05.484022+0000","last_active":"2026-03-10T11:32:13.752919+0000","last_peered":"2026-03-10T11:32:13.752919+0000","last_clean":"2026-03-10T11:32:13.752919+0000","last_became_active":"2026-03-10T11:32:05.483921+0000","last_became_peered":"2026-03-10T11:32:05.483921+0000","last_unstale":"2026-03-10T11:32:13.752919+0000","last_undegraded":"2026-03-10T11:32:13.752919+0000","last_fullsized":"2026-03-10T11:32:13.752919+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:52:04.652862+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":92,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":10,"num_read_kb":10,"num_write":4,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,0],"acting":[7,4,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.9","version":"55'8","reported_seq":30,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752990+0000","last_change":"2026-03-10T11:32:09.531130+0000","last_active":"2026-03-10T11:32:13.752990+0000","last_peered":"2026-03-10T11:32:13.752990+0000","last_clean":"2026-03-10T11:32:13.752990+0000","last_became_active":"2026-03-10T11:32:09.530942+0000","last_became_peered":"2026-03-10T11:32:09.530942+0000","last_unstale":"2026-03-10T11:32:13.752990+0000","last_undegraded":"2026-03-10T11:32:13.752990+0000","last_fullsized":"2026-03-10T11:32:13.752990+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:21:44.911594+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,6,4],"acting":[7,6,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"6.a","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434856+0000","last_change":"2026-03-10T11:32:11.573996+0000","last_active":"2026-03-10T11:32:13.434856+0000","last_peered":"2026-03-10T11:32:13.434856+0000","last_clean":"2026-03-10T11:32:13.434856+0000","last_became_active":"2026-03-10T11:32:11.573890+0000","last_became_peered":"2026-03-10T11:32:11.573890+0000","last_unstale":"2026-03-10T11:32:13.434856+0000","last_undegraded":"2026-03-10T11:32:13.434856+0000","last_fullsized":"2026-03-10T11:32:13.434856+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:55:47.737007+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,6,0],"acting":[5,6,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"3.10","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431436+0000","last_change":"2026-03-10T11:32:05.480882+0000","last_active":"2026-03-10T11:32:13.431436+0000","last_peered":"2026-03-10T11:32:13.431436+0000","last_clean":"2026-03-10T11:32:13.431436+0000","last_became_active":"2026-03-10T11:32:05.480785+0000","last_became_peered":"2026-03-10T11:32:05.480785+0000","last_unstale":"2026-03-10T11:32:13.431436+0000","last_undegraded":"2026-03-10T11:32:13.431436+0000","last_fullsized":"2026-03-10T11:32:13.431436+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:58:13.892454+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,0,5],"acting":[6,0,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"4.17","version":"55'6","reported_seq":32,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432998+0000","last_change":"2026-03-10T11:32:07.598865+0000","last_active":"2026-03-10T11:32:13.432998+0000","last_peered":"2026-03-10T11:32:13.432998+0000","last_clean":"2026-03-10T11:32:13.432998+0000","last_became_active":"2026-03-10T11:32:07.598687+0000","last_became_peered":"2026-03-10T11:32:07.598687+0000","last_unstale":"2026-03-10T11:32:13.432998+0000","last_undegraded":"2026-03-10T11:32:13.432998+0000","last_fullsized":"2026-03-10T11:32:13.432998+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":6,"log_dups_size":0,"ondisk_log_size":6,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:45:00.774145+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":3,"num_object_clones":0,"num_object_copies":9,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":3,"num_whiteouts":0,"num_read":9,"num_read_kb":6,"num_write":6,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,1],"acting":[3,7,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.16","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435693+0000","last_change":"2026-03-10T11:32:09.507650+0000","last_active":"2026-03-10T11:32:13.435693+0000","last_peered":"2026-03-10T11:32:13.435693+0000","last_clean":"2026-03-10T11:32:13.435693+0000","last_became_active":"2026-03-10T11:32:09.501191+0000","last_became_peered":"2026-03-10T11:32:09.501191+0000","last_unstale":"2026-03-10T11:32:13.435693+0000","last_undegraded":"2026-03-10T11:32:13.435693+0000","last_fullsized":"2026-03-10T11:32:13.435693+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:30:26.718870+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,3,1],"acting":[5,3,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.15","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753513+0000","last_change":"2026-03-10T11:32:11.577353+0000","last_active":"2026-03-10T11:32:13.753513+0000","last_peered":"2026-03-10T11:32:13.753513+0000","last_clean":"2026-03-10T11:32:13.753513+0000","last_became_active":"2026-03-10T11:32:11.575570+0000","last_became_peered":"2026-03-10T11:32:11.575570+0000","last_unstale":"2026-03-10T11:32:13.753513+0000","last_undegraded":"2026-03-10T11:32:13.753513+0000","last_fullsized":"2026-03-10T11:32:13.753513+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:52:18.344523+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,6,4],"acting":[7,6,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"4.16","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539668+0000","last_change":"2026-03-10T11:32:07.597957+0000","last_active":"2026-03-10T11:32:13.539668+0000","last_peered":"2026-03-10T11:32:13.539668+0000","last_clean":"2026-03-10T11:32:13.539668+0000","last_became_active":"2026-03-10T11:32:07.597610+0000","last_became_peered":"2026-03-10T11:32:07.597610+0000","last_unstale":"2026-03-10T11:32:13.539668+0000","last_undegraded":"2026-03-10T11:32:13.539668+0000","last_fullsized":"2026-03-10T11:32:13.539668+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:44:45.379060+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,3,7],"acting":[0,3,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"3.11","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752279+0000","last_change":"2026-03-10T11:32:05.479640+0000","last_active":"2026-03-10T11:32:13.752279+0000","last_peered":"2026-03-10T11:32:13.752279+0000","last_clean":"2026-03-10T11:32:13.752279+0000","last_became_active":"2026-03-10T11:32:05.479282+0000","last_became_peered":"2026-03-10T11:32:05.479282+0000","last_unstale":"2026-03-10T11:32:13.752279+0000","last_undegraded":"2026-03-10T11:32:13.752279+0000","last_fullsized":"2026-03-10T11:32:13.752279+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:25:22.717010+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,6],"acting":[7,4,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.17","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432008+0000","last_change":"2026-03-10T11:32:09.503165+0000","last_active":"2026-03-10T11:32:13.432008+0000","last_peered":"2026-03-10T11:32:13.432008+0000","last_clean":"2026-03-10T11:32:13.432008+0000","last_became_active":"2026-03-10T11:32:09.503019+0000","last_became_peered":"2026-03-10T11:32:09.503019+0000","last_unstale":"2026-03-10T11:32:13.432008+0000","last_undegraded":"2026-03-10T11:32:13.432008+0000","last_fullsized":"2026-03-10T11:32:13.432008+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:58:40.889689+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,7],"acting":[3,1,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"6.14","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.094844+0000","last_change":"2026-03-10T11:32:11.508463+0000","last_active":"2026-03-10T11:32:14.094844+0000","last_peered":"2026-03-10T11:32:14.094844+0000","last_clean":"2026-03-10T11:32:14.094844+0000","last_became_active":"2026-03-10T11:32:11.508373+0000","last_became_peered":"2026-03-10T11:32:11.508373+0000","last_unstale":"2026-03-10T11:32:14.094844+0000","last_undegraded":"2026-03-10T11:32:14.094844+0000","last_fullsized":"2026-03-10T11:32:14.094844+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:11:49.345325+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,4,7],"acting":[2,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"4.15","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435565+0000","last_change":"2026-03-10T11:32:07.597830+0000","last_active":"2026-03-10T11:32:13.435565+0000","last_peered":"2026-03-10T11:32:13.435565+0000","last_clean":"2026-03-10T11:32:13.435565+0000","last_became_active":"2026-03-10T11:32:07.597705+0000","last_became_peered":"2026-03-10T11:32:07.597705+0000","last_unstale":"2026-03-10T11:32:13.435565+0000","last_undegraded":"2026-03-10T11:32:13.435565+0000","last_fullsized":"2026-03-10T11:32:13.435565+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:51:57.119228+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,7,3],"acting":[5,7,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"3.12","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.538926+0000","last_change":"2026-03-10T11:32:05.494103+0000","last_active":"2026-03-10T11:32:13.538926+0000","last_peered":"2026-03-10T11:32:13.538926+0000","last_clean":"2026-03-10T11:32:13.538926+0000","last_became_active":"2026-03-10T11:32:05.493873+0000","last_became_peered":"2026-03-10T11:32:05.493873+0000","last_unstale":"2026-03-10T11:32:13.538926+0000","last_undegraded":"2026-03-10T11:32:13.538926+0000","last_fullsized":"2026-03-10T11:32:13.538926+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:30:14.014213+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,7,3],"acting":[0,7,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"5.14","version":"55'8","reported_seq":29,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432754+0000","last_change":"2026-03-10T11:32:09.503097+0000","last_active":"2026-03-10T11:32:13.432754+0000","last_peered":"2026-03-10T11:32:13.432754+0000","last_clean":"2026-03-10T11:32:13.432754+0000","last_became_active":"2026-03-10T11:32:09.502897+0000","last_became_peered":"2026-03-10T11:32:09.502897+0000","last_unstale":"2026-03-10T11:32:13.432754+0000","last_undegraded":"2026-03-10T11:32:13.432754+0000","last_fullsized":"2026-03-10T11:32:13.432754+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:19:50.521963+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,2],"acting":[3,7,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"6.17","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058501+0000","last_change":"2026-03-10T11:32:11.508381+0000","last_active":"2026-03-10T11:32:14.058501+0000","last_peered":"2026-03-10T11:32:14.058501+0000","last_clean":"2026-03-10T11:32:14.058501+0000","last_became_active":"2026-03-10T11:32:11.508250+0000","last_became_peered":"2026-03-10T11:32:11.508250+0000","last_unstale":"2026-03-10T11:32:14.058501+0000","last_undegraded":"2026-03-10T11:32:14.058501+0000","last_fullsized":"2026-03-10T11:32:14.058501+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:44:59.324849+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,2,5],"acting":[4,2,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"4.14","version":"55'10","reported_seq":38,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432952+0000","last_change":"2026-03-10T11:32:07.598047+0000","last_active":"2026-03-10T11:32:13.432952+0000","last_peered":"2026-03-10T11:32:13.432952+0000","last_clean":"2026-03-10T11:32:13.432952+0000","last_became_active":"2026-03-10T11:32:07.596958+0000","last_became_peered":"2026-03-10T11:32:07.596958+0000","last_unstale":"2026-03-10T11:32:13.432952+0000","last_undegraded":"2026-03-10T11:32:13.432952+0000","last_fullsized":"2026-03-10T11:32:13.432952+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":10,"log_dups_size":0,"ondisk_log_size":10,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:41:42.255637+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":15,"num_read_kb":10,"num_write":10,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,7],"acting":[3,1,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.13","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752816+0000","last_change":"2026-03-10T11:32:05.480017+0000","last_active":"2026-03-10T11:32:13.752816+0000","last_peered":"2026-03-10T11:32:13.752816+0000","last_clean":"2026-03-10T11:32:13.752816+0000","last_became_active":"2026-03-10T11:32:05.479411+0000","last_became_peered":"2026-03-10T11:32:05.479411+0000","last_unstale":"2026-03-10T11:32:13.752816+0000","last_undegraded":"2026-03-10T11:32:13.752816+0000","last_fullsized":"2026-03-10T11:32:13.752816+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:31:09.439155+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,2],"acting":[7,4,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.15","version":"55'8","reported_seq":32,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435785+0000","last_change":"2026-03-10T11:32:09.508210+0000","last_active":"2026-03-10T11:32:13.435785+0000","last_peered":"2026-03-10T11:32:13.435785+0000","last_clean":"2026-03-10T11:32:13.435785+0000","last_became_active":"2026-03-10T11:32:09.508106+0000","last_became_peered":"2026-03-10T11:32:09.508106+0000","last_unstale":"2026-03-10T11:32:13.435785+0000","last_undegraded":"2026-03-10T11:32:13.435785+0000","last_fullsized":"2026-03-10T11:32:13.435785+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:26:43.229572+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1,0],"acting":[5,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.16","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539060+0000","last_change":"2026-03-10T11:32:11.506003+0000","last_active":"2026-03-10T11:32:13.539060+0000","last_peered":"2026-03-10T11:32:13.539060+0000","last_clean":"2026-03-10T11:32:13.539060+0000","last_became_active":"2026-03-10T11:32:11.505923+0000","last_became_peered":"2026-03-10T11:32:11.505923+0000","last_unstale":"2026-03-10T11:32:13.539060+0000","last_undegraded":"2026-03-10T11:32:13.539060+0000","last_fullsized":"2026-03-10T11:32:13.539060+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:55:53.821261+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,7,3],"acting":[0,7,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.13","version":"55'11","reported_seq":42,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059558+0000","last_change":"2026-03-10T11:32:07.480647+0000","last_active":"2026-03-10T11:32:14.059558+0000","last_peered":"2026-03-10T11:32:14.059558+0000","last_clean":"2026-03-10T11:32:14.059558+0000","last_became_active":"2026-03-10T11:32:07.480488+0000","last_became_peered":"2026-03-10T11:32:07.480488+0000","last_unstale":"2026-03-10T11:32:14.059558+0000","last_undegraded":"2026-03-10T11:32:14.059558+0000","last_fullsized":"2026-03-10T11:32:14.059558+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":11,"log_dups_size":0,"ondisk_log_size":11,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:58:00.369153+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":20,"num_read_kb":13,"num_write":12,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,6,1],"acting":[4,6,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.14","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058329+0000","last_change":"2026-03-10T11:32:05.492719+0000","last_active":"2026-03-10T11:32:14.058329+0000","last_peered":"2026-03-10T11:32:14.058329+0000","last_clean":"2026-03-10T11:32:14.058329+0000","last_became_active":"2026-03-10T11:32:05.492463+0000","last_became_peered":"2026-03-10T11:32:05.492463+0000","last_unstale":"2026-03-10T11:32:14.058329+0000","last_undegraded":"2026-03-10T11:32:14.058329+0000","last_fullsized":"2026-03-10T11:32:14.058329+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:29:31.841348+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,7,6],"acting":[4,7,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"5.12","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757427+0000","last_change":"2026-03-10T11:32:09.511793+0000","last_active":"2026-03-10T11:32:13.757427+0000","last_peered":"2026-03-10T11:32:13.757427+0000","last_clean":"2026-03-10T11:32:13.757427+0000","last_became_active":"2026-03-10T11:32:09.509850+0000","last_became_peered":"2026-03-10T11:32:09.509850+0000","last_unstale":"2026-03-10T11:32:13.757427+0000","last_undegraded":"2026-03-10T11:32:13.757427+0000","last_fullsized":"2026-03-10T11:32:13.757427+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:23:58.904271+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,3],"acting":[1,5,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"6.11","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431948+0000","last_change":"2026-03-10T11:32:11.514919+0000","last_active":"2026-03-10T11:32:13.431948+0000","last_peered":"2026-03-10T11:32:13.431948+0000","last_clean":"2026-03-10T11:32:13.431948+0000","last_became_active":"2026-03-10T11:32:11.514837+0000","last_became_peered":"2026-03-10T11:32:11.514837+0000","last_unstale":"2026-03-10T11:32:13.431948+0000","last_undegraded":"2026-03-10T11:32:13.431948+0000","last_fullsized":"2026-03-10T11:32:13.431948+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:27:53.843660+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0,5],"acting":[3,0,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.12","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757383+0000","last_change":"2026-03-10T11:32:07.482374+0000","last_active":"2026-03-10T11:32:13.757383+0000","last_peered":"2026-03-10T11:32:13.757383+0000","last_clean":"2026-03-10T11:32:13.757383+0000","last_became_active":"2026-03-10T11:32:07.482274+0000","last_became_peered":"2026-03-10T11:32:07.482274+0000","last_unstale":"2026-03-10T11:32:13.757383+0000","last_undegraded":"2026-03-10T11:32:13.757383+0000","last_fullsized":"2026-03-10T11:32:13.757383+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:41:26.092888+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,6,2],"acting":[1,6,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.15","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752290+0000","last_change":"2026-03-10T11:32:05.480988+0000","last_active":"2026-03-10T11:32:13.752290+0000","last_peered":"2026-03-10T11:32:13.752290+0000","last_clean":"2026-03-10T11:32:13.752290+0000","last_became_active":"2026-03-10T11:32:05.480882+0000","last_became_peered":"2026-03-10T11:32:05.480882+0000","last_unstale":"2026-03-10T11:32:13.752290+0000","last_undegraded":"2026-03-10T11:32:13.752290+0000","last_fullsized":"2026-03-10T11:32:13.752290+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:12:29.005725+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,3,4],"acting":[7,3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.13","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432116+0000","last_change":"2026-03-10T11:32:09.498920+0000","last_active":"2026-03-10T11:32:13.432116+0000","last_peered":"2026-03-10T11:32:13.432116+0000","last_clean":"2026-03-10T11:32:13.432116+0000","last_became_active":"2026-03-10T11:32:09.498815+0000","last_became_peered":"2026-03-10T11:32:09.498815+0000","last_unstale":"2026-03-10T11:32:13.432116+0000","last_undegraded":"2026-03-10T11:32:13.432116+0000","last_fullsized":"2026-03-10T11:32:13.432116+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:08:28.641280+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0,1],"acting":[3,0,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"6.10","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539907+0000","last_change":"2026-03-10T11:32:11.510088+0000","last_active":"2026-03-10T11:32:13.539907+0000","last_peered":"2026-03-10T11:32:13.539907+0000","last_clean":"2026-03-10T11:32:13.539907+0000","last_became_active":"2026-03-10T11:32:11.510003+0000","last_became_peered":"2026-03-10T11:32:11.510003+0000","last_unstale":"2026-03-10T11:32:13.539907+0000","last_undegraded":"2026-03-10T11:32:13.539907+0000","last_fullsized":"2026-03-10T11:32:13.539907+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:45:24.974948+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,5,1],"acting":[0,5,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.11","version":"55'11","reported_seq":42,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432908+0000","last_change":"2026-03-10T11:32:07.598238+0000","last_active":"2026-03-10T11:32:13.432908+0000","last_peered":"2026-03-10T11:32:13.432908+0000","last_clean":"2026-03-10T11:32:13.432908+0000","last_became_active":"2026-03-10T11:32:07.598143+0000","last_became_peered":"2026-03-10T11:32:07.598143+0000","last_unstale":"2026-03-10T11:32:13.432908+0000","last_undegraded":"2026-03-10T11:32:13.432908+0000","last_fullsized":"2026-03-10T11:32:13.432908+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":11,"log_dups_size":0,"ondisk_log_size":11,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:11:32.909846+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":20,"num_read_kb":13,"num_write":12,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,6],"acting":[3,7,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.16","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435324+0000","last_change":"2026-03-10T11:32:05.472568+0000","last_active":"2026-03-10T11:32:13.435324+0000","last_peered":"2026-03-10T11:32:13.435324+0000","last_clean":"2026-03-10T11:32:13.435324+0000","last_became_active":"2026-03-10T11:32:05.472437+0000","last_became_peered":"2026-03-10T11:32:05.472437+0000","last_unstale":"2026-03-10T11:32:13.435324+0000","last_undegraded":"2026-03-10T11:32:13.435324+0000","last_fullsized":"2026-03-10T11:32:13.435324+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:52:02.801467+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,7,1],"acting":[5,7,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"5.10","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752768+0000","last_change":"2026-03-10T11:32:09.519265+0000","last_active":"2026-03-10T11:32:13.752768+0000","last_peered":"2026-03-10T11:32:13.752768+0000","last_clean":"2026-03-10T11:32:13.752768+0000","last_became_active":"2026-03-10T11:32:09.519124+0000","last_became_peered":"2026-03-10T11:32:09.519124+0000","last_unstale":"2026-03-10T11:32:13.752768+0000","last_undegraded":"2026-03-10T11:32:13.752768+0000","last_fullsized":"2026-03-10T11:32:13.752768+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:26:19.264009+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,6],"acting":[7,4,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"6.13","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432894+0000","last_change":"2026-03-10T11:32:11.573852+0000","last_active":"2026-03-10T11:32:13.432894+0000","last_peered":"2026-03-10T11:32:13.432894+0000","last_clean":"2026-03-10T11:32:13.432894+0000","last_became_active":"2026-03-10T11:32:11.573716+0000","last_became_peered":"2026-03-10T11:32:11.573716+0000","last_unstale":"2026-03-10T11:32:13.432894+0000","last_undegraded":"2026-03-10T11:32:13.432894+0000","last_fullsized":"2026-03-10T11:32:13.432894+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:33:18.220367+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0,6],"acting":[3,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.10","version":"55'4","reported_seq":29,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432725+0000","last_change":"2026-03-10T11:32:07.493099+0000","last_active":"2026-03-10T11:32:13.432725+0000","last_peered":"2026-03-10T11:32:13.432725+0000","last_clean":"2026-03-10T11:32:13.432725+0000","last_became_active":"2026-03-10T11:32:07.493016+0000","last_became_peered":"2026-03-10T11:32:07.493016+0000","last_unstale":"2026-03-10T11:32:13.432725+0000","last_undegraded":"2026-03-10T11:32:13.432725+0000","last_fullsized":"2026-03-10T11:32:13.432725+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":4,"log_dups_size":0,"ondisk_log_size":4,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:32:04.890080+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":6,"num_read_kb":4,"num_write":4,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,6],"acting":[3,1,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.17","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.538956+0000","last_change":"2026-03-10T11:32:05.481179+0000","last_active":"2026-03-10T11:32:13.538956+0000","last_peered":"2026-03-10T11:32:13.538956+0000","last_clean":"2026-03-10T11:32:13.538956+0000","last_became_active":"2026-03-10T11:32:05.481078+0000","last_became_peered":"2026-03-10T11:32:05.481078+0000","last_unstale":"2026-03-10T11:32:13.538956+0000","last_undegraded":"2026-03-10T11:32:13.538956+0000","last_fullsized":"2026-03-10T11:32:13.538956+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:20:19.773427+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,5,3],"acting":[0,5,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"5.11","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431859+0000","last_change":"2026-03-10T11:32:09.517608+0000","last_active":"2026-03-10T11:32:13.431859+0000","last_peered":"2026-03-10T11:32:13.431859+0000","last_clean":"2026-03-10T11:32:13.431859+0000","last_became_active":"2026-03-10T11:32:09.517388+0000","last_became_peered":"2026-03-10T11:32:09.517388+0000","last_unstale":"2026-03-10T11:32:13.431859+0000","last_undegraded":"2026-03-10T11:32:13.431859+0000","last_fullsized":"2026-03-10T11:32:13.431859+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:30:58.947416+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,4,7],"acting":[6,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"6.12","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752434+0000","last_change":"2026-03-10T11:32:11.509710+0000","last_active":"2026-03-10T11:32:13.752434+0000","last_peered":"2026-03-10T11:32:13.752434+0000","last_clean":"2026-03-10T11:32:13.752434+0000","last_became_active":"2026-03-10T11:32:11.508854+0000","last_became_peered":"2026-03-10T11:32:11.508854+0000","last_unstale":"2026-03-10T11:32:13.752434+0000","last_undegraded":"2026-03-10T11:32:13.752434+0000","last_fullsized":"2026-03-10T11:32:13.752434+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:36:56.911178+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,2,4],"acting":[7,2,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"6.1d","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.756579+0000","last_change":"2026-03-10T11:32:11.509797+0000","last_active":"2026-03-10T11:32:13.756579+0000","last_peered":"2026-03-10T11:32:13.756579+0000","last_clean":"2026-03-10T11:32:13.756579+0000","last_became_active":"2026-03-10T11:32:11.509718+0000","last_became_peered":"2026-03-10T11:32:11.509718+0000","last_unstale":"2026-03-10T11:32:13.756579+0000","last_undegraded":"2026-03-10T11:32:13.756579+0000","last_fullsized":"2026-03-10T11:32:13.756579+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:40:19.756646+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,4],"acting":[1,5,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.18","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432542+0000","last_change":"2026-03-10T11:32:05.488107+0000","last_active":"2026-03-10T11:32:13.432542+0000","last_peered":"2026-03-10T11:32:13.432542+0000","last_clean":"2026-03-10T11:32:13.432542+0000","last_became_active":"2026-03-10T11:32:05.487957+0000","last_became_peered":"2026-03-10T11:32:05.487957+0000","last_unstale":"2026-03-10T11:32:13.432542+0000","last_undegraded":"2026-03-10T11:32:13.432542+0000","last_fullsized":"2026-03-10T11:32:13.432542+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:30:30.807507+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0,1],"acting":[3,0,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.1f","version":"55'11","reported_seq":42,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431653+0000","last_change":"2026-03-10T11:32:07.492632+0000","last_active":"2026-03-10T11:32:13.431653+0000","last_peered":"2026-03-10T11:32:13.431653+0000","last_clean":"2026-03-10T11:32:13.431653+0000","last_became_active":"2026-03-10T11:32:07.492509+0000","last_became_peered":"2026-03-10T11:32:07.492509+0000","last_unstale":"2026-03-10T11:32:13.431653+0000","last_undegraded":"2026-03-10T11:32:13.431653+0000","last_fullsized":"2026-03-10T11:32:13.431653+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":11,"log_dups_size":0,"ondisk_log_size":11,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:55:48.589620+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":20,"num_read_kb":13,"num_write":12,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,5,1],"acting":[6,5,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"5.1e","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539543+0000","last_change":"2026-03-10T11:32:09.505710+0000","last_active":"2026-03-10T11:32:13.539543+0000","last_peered":"2026-03-10T11:32:13.539543+0000","last_clean":"2026-03-10T11:32:13.539543+0000","last_became_active":"2026-03-10T11:32:09.503635+0000","last_became_peered":"2026-03-10T11:32:09.503635+0000","last_unstale":"2026-03-10T11:32:13.539543+0000","last_undegraded":"2026-03-10T11:32:13.539543+0000","last_fullsized":"2026-03-10T11:32:13.539543+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:44:14.135459+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,7,2],"acting":[0,7,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]}],"pool_stats":[{"poolid":6,"num_pg":32,"stat_sum":{"num_bytes":416,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":3,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1248,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":2,"ondisk_log_size":2,"up":96,"acting":96,"num_store_stats":8},{"poolid":5,"num_pg":32,"stat_sum":{"num_bytes":0,"num_objects":8,"num_object_clones":0,"num_object_copies":24,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":8,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":64,"ondisk_log_size":64,"up":96,"acting":96,"num_store_stats":8},{"poolid":4,"num_pg":32,"stat_sum":{"num_bytes":3702,"num_objects":178,"num_object_clones":0,"num_object_copies":534,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":178,"num_whiteouts":0,"num_read":698,"num_read_kb":455,"num_write":417,"num_write_kb":34,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":417792,"data_stored":11106,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":393,"ondisk_log_size":393,"up":96,"acting":96,"num_store_stats":8},{"poolid":3,"num_pg":32,"stat_sum":{"num_bytes":1613,"num_objects":6,"num_object_clones":0,"num_object_copies":18,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":6,"num_whiteouts":0,"num_read":24,"num_read_kb":24,"num_write":10,"num_write_kb":6,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":73728,"data_stored":4839,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":6,"ondisk_log_size":6,"up":96,"acting":96,"num_store_stats":8},{"poolid":2,"num_pg":3,"stat_sum":{"num_bytes":408,"num_objects":3,"num_object_clones":0,"num_object_copies":9,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":3,"num_whiteouts":0,"num_read":8,"num_read_kb":3,"num_write":6,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1224,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":8,"ondisk_log_size":8,"up":9,"acting":9,"num_store_stats":7},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":459280,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2777088,"data_stored":2755680,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":7}],"osd_stats":[{"osd":7,"up_from":44,"seq":188978561029,"num_pgs":53,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27904,"kb_used_data":1072,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939520,"statfs":{"total":21470642176,"available":21442068480,"internally_reserved":0,"allocated":1097728,"data_stored":719765,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":39,"seq":167503724551,"num_pgs":43,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27888,"kb_used_data":1052,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939536,"statfs":{"total":21470642176,"available":21442084864,"internally_reserved":0,"allocated":1077248,"data_stored":718123,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":32,"seq":137438953481,"num_pgs":47,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27444,"kb_used_data":604,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939980,"statfs":{"total":21470642176,"available":21442539520,"internally_reserved":0,"allocated":618496,"data_stored":258779,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":27,"seq":115964117003,"num_pgs":58,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27480,"kb_used_data":632,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939944,"statfs":{"total":21470642176,"available":21442502656,"internally_reserved":0,"allocated":647168,"data_stored":260441,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":26,"apply_latency_ms":26,"commit_latency_ns":26000000,"apply_latency_ns":26000000},"alerts":[]},{"osd":3,"up_from":22,"seq":94489280525,"num_pgs":56,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27472,"kb_used_data":636,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939952,"statfs":{"total":21470642176,"available":21442510848,"internally_reserved":0,"allocated":651264,"data_stored":259311,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":5,"apply_latency_ms":5,"commit_latency_ns":5000000,"apply_latency_ns":5000000},"alerts":[]},{"osd":2,"up_from":16,"seq":68719476751,"num_pgs":36,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27436,"kb_used_data":596,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939988,"statfs":{"total":21470642176,"available":21442547712,"internally_reserved":0,"allocated":610304,"data_stored":259548,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":2,"apply_latency_ms":2,"commit_latency_ns":2000000,"apply_latency_ns":2000000},"alerts":[]},{"osd":1,"up_from":12,"seq":51539607569,"num_pgs":57,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27492,"kb_used_data":652,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939932,"statfs":{"total":21470642176,"available":21442490368,"internally_reserved":0,"allocated":667648,"data_stored":252896,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738388,"num_pgs":46,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27912,"kb_used_data":1072,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939512,"statfs":{"total":21470642176,"available":21442060288,"internally_reserved":0,"allocated":1097728,"data_stored":718481,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":408,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":12288,"data_stored":138,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":12288,"data_stored":528,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":993,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":46,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1613,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":436,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":12288,"data_stored":1085,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":49152,"data_stored":1320,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":90112,"data_stored":2338,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":32768,"data_stored":798,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":73728,"data_stored":1898,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":53248,"data_stored":1474,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":36864,"data_stored":990,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":36864,"data_stored":1034,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":45056,"data_stored":1254,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":13,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":403,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":13,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":416,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":403,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T11:32:20.290 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph pg dump --format=json 2026-03-10T11:32:20.480 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 podman[79486]: 2026-03-10 11:32:20.146365163 +0000 UTC m=+0.018614834 container create b914580585251e79c1fb0ea6ce1c4175355c80a6d1df2396e806728636ee5251 (image=quay.io/prometheus/prometheus:v2.51.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:20.480 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 podman[79486]: 2026-03-10 11:32:20.178170605 +0000 UTC m=+0.050420266 container init b914580585251e79c1fb0ea6ce1c4175355c80a6d1df2396e806728636ee5251 (image=quay.io/prometheus/prometheus:v2.51.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:20.480 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 podman[79486]: 2026-03-10 11:32:20.180709718 +0000 UTC m=+0.052959379 container start b914580585251e79c1fb0ea6ce1c4175355c80a6d1df2396e806728636ee5251 (image=quay.io/prometheus/prometheus:v2.51.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:20.480 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 bash[79486]: b914580585251e79c1fb0ea6ce1c4175355c80a6d1df2396e806728636ee5251 2026-03-10T11:32:20.480 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 podman[79486]: 2026-03-10 11:32:20.138613955 +0000 UTC m=+0.010863626 image pull 1d3b7f56885b6dd623f1785be963aa9c195f86bc256ea454e8d02a7980b79c53 quay.io/prometheus/prometheus:v2.51.0 2026-03-10T11:32:20.480 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 systemd[1]: Started Ceph prometheus.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.213Z caller=main.go:617 level=info msg="Starting Prometheus Server" mode=server version="(version=2.51.0, branch=HEAD, revision=c05c15512acb675e3f6cd662a6727854e93fc024)" 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.213Z caller=main.go:622 level=info build_context="(go=go1.22.1, platform=linux/amd64, user=root@b5723e458358, date=20240319-10:54:45, tags=netgo,builtinassets,stringlabels)" 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.213Z caller=main.go:623 level=info host_details="(Linux 5.14.0-686.el9.x86_64 #1 SMP PREEMPT_DYNAMIC Thu Feb 19 10:49:27 UTC 2026 x86_64 vm09 (none))" 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.213Z caller=main.go:624 level=info fd_limits="(soft=1048576, hard=1048576)" 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.213Z caller=main.go:625 level=info vm_limits="(soft=unlimited, hard=unlimited)" 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.217Z caller=web.go:568 level=info component=web msg="Start listening for connections" address=:9095 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.217Z caller=main.go:1129 level=info msg="Starting TSDB ..." 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.219Z caller=tls_config.go:313 level=info component=web msg="Listening on" address=[::]:9095 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.219Z caller=tls_config.go:316 level=info component=web msg="TLS is disabled." http2=false address=[::]:9095 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.221Z caller=head.go:616 level=info component=tsdb msg="Replaying on-disk memory mappable chunks if any" 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.221Z caller=head.go:698 level=info component=tsdb msg="On-disk memory mappable chunks replay completed" duration=1.993µs 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.221Z caller=head.go:706 level=info component=tsdb msg="Replaying WAL, this may take a while" 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.221Z caller=head.go:778 level=info component=tsdb msg="WAL segment loaded" segment=0 maxSegment=0 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.221Z caller=head.go:815 level=info component=tsdb msg="WAL replay completed" checkpoint_replay_duration=20.398µs wal_replay_duration=114.986µs wbl_replay_duration=330ns total_replay_duration=191.227µs 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.222Z caller=main.go:1150 level=info fs_type=XFS_SUPER_MAGIC 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.222Z caller=main.go:1153 level=info msg="TSDB started" 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.222Z caller=main.go:1335 level=info msg="Loading configuration file" filename=/etc/prometheus/prometheus.yml 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.235Z caller=main.go:1372 level=info msg="Completed loading of configuration file" filename=/etc/prometheus/prometheus.yml totalDuration=13.397557ms db_storage=771ns remote_storage=1.041µs web_handler=281ns query_engine=581ns scrape=1.054274ms scrape_sd=67.116µs notify=561ns notify_sd=652ns rules=12.026812ms tracing=6.361µs 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.235Z caller=main.go:1114 level=info msg="Server is ready to receive web requests." 2026-03-10T11:32:20.481 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:20.235Z caller=manager.go:163 level=info component="rule manager" msg="Starting rule manager..." 2026-03-10T11:32:20.511 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:20.743 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:32:20.748 INFO:teuthology.orchestra.run.vm06.stderr:dumped all 2026-03-10T11:32:20.823 INFO:teuthology.orchestra.run.vm06.stdout:{"pg_ready":true,"pg_map":{"version":112,"stamp":"2026-03-10T11:32:20.277655+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":465419,"num_objects":199,"num_object_clones":0,"num_object_copies":597,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":199,"num_whiteouts":0,"num_read":776,"num_read_kb":519,"num_write":493,"num_write_kb":629,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":459280,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":505,"ondisk_log_size":505,"up":396,"acting":396,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":396,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":8,"kb":167739392,"kb_used":221040,"kb_used_data":6332,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167518352,"statfs":{"total":171765137408,"available":171538792448,"internally_reserved":0,"allocated":6483968,"data_stored":3455089,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12711,"internal_metadata":219663961},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":4353,"num_objects":186,"num_object_clones":0,"num_object_copies":558,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":186,"num_whiteouts":0,"num_read":709,"num_read_kb":465,"num_write":423,"num_write_kb":36,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"8.001757"},"pg_stats":[{"pgid":"3.1f","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.538855+0000","last_change":"2026-03-10T11:32:05.482279+0000","last_active":"2026-03-10T11:32:13.538855+0000","last_peered":"2026-03-10T11:32:13.538855+0000","last_clean":"2026-03-10T11:32:13.538855+0000","last_became_active":"2026-03-10T11:32:05.482189+0000","last_became_peered":"2026-03-10T11:32:05.482189+0000","last_unstale":"2026-03-10T11:32:13.538855+0000","last_undegraded":"2026-03-10T11:32:13.538855+0000","last_fullsized":"2026-03-10T11:32:13.538855+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:33:08.999865+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,5,2],"acting":[0,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.18","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059914+0000","last_change":"2026-03-10T11:32:07.478618+0000","last_active":"2026-03-10T11:32:14.059914+0000","last_peered":"2026-03-10T11:32:14.059914+0000","last_clean":"2026-03-10T11:32:14.059914+0000","last_became_active":"2026-03-10T11:32:07.478521+0000","last_became_peered":"2026-03-10T11:32:07.478521+0000","last_unstale":"2026-03-10T11:32:14.059914+0000","last_undegraded":"2026-03-10T11:32:14.059914+0000","last_fullsized":"2026-03-10T11:32:14.059914+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:31:56.427965+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,6,1],"acting":[4,6,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"5.19","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757486+0000","last_change":"2026-03-10T11:32:09.511642+0000","last_active":"2026-03-10T11:32:13.757486+0000","last_peered":"2026-03-10T11:32:13.757486+0000","last_clean":"2026-03-10T11:32:13.757486+0000","last_became_active":"2026-03-10T11:32:09.509697+0000","last_became_peered":"2026-03-10T11:32:09.509697+0000","last_unstale":"2026-03-10T11:32:13.757486+0000","last_undegraded":"2026-03-10T11:32:13.757486+0000","last_fullsized":"2026-03-10T11:32:13.757486+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:00:25.667302+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,7],"acting":[1,5,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"6.1a","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059887+0000","last_change":"2026-03-10T11:32:11.509684+0000","last_active":"2026-03-10T11:32:14.059887+0000","last_peered":"2026-03-10T11:32:14.059887+0000","last_clean":"2026-03-10T11:32:14.059887+0000","last_became_active":"2026-03-10T11:32:11.509561+0000","last_became_peered":"2026-03-10T11:32:11.509561+0000","last_unstale":"2026-03-10T11:32:14.059887+0000","last_undegraded":"2026-03-10T11:32:14.059887+0000","last_fullsized":"2026-03-10T11:32:14.059887+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:03:55.104524+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,5,1],"acting":[4,5,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"6.1b","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432284+0000","last_change":"2026-03-10T11:32:11.573329+0000","last_active":"2026-03-10T11:32:13.432284+0000","last_peered":"2026-03-10T11:32:13.432284+0000","last_clean":"2026-03-10T11:32:13.432284+0000","last_became_active":"2026-03-10T11:32:11.573182+0000","last_became_peered":"2026-03-10T11:32:11.573182+0000","last_unstale":"2026-03-10T11:32:13.432284+0000","last_undegraded":"2026-03-10T11:32:13.432284+0000","last_fullsized":"2026-03-10T11:32:13.432284+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:50:42.681652+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,6],"acting":[3,7,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.1e","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432298+0000","last_change":"2026-03-10T11:32:05.482297+0000","last_active":"2026-03-10T11:32:13.432298+0000","last_peered":"2026-03-10T11:32:13.432298+0000","last_clean":"2026-03-10T11:32:13.432298+0000","last_became_active":"2026-03-10T11:32:05.482171+0000","last_became_peered":"2026-03-10T11:32:05.482171+0000","last_unstale":"2026-03-10T11:32:13.432298+0000","last_undegraded":"2026-03-10T11:32:13.432298+0000","last_fullsized":"2026-03-10T11:32:13.432298+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:02:00.446821+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,6,2],"acting":[3,6,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.19","version":"55'15","reported_seq":48,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432324+0000","last_change":"2026-03-10T11:32:07.482046+0000","last_active":"2026-03-10T11:32:13.432324+0000","last_peered":"2026-03-10T11:32:13.432324+0000","last_clean":"2026-03-10T11:32:13.432324+0000","last_became_active":"2026-03-10T11:32:07.481957+0000","last_became_peered":"2026-03-10T11:32:07.481957+0000","last_unstale":"2026-03-10T11:32:13.432324+0000","last_undegraded":"2026-03-10T11:32:13.432324+0000","last_fullsized":"2026-03-10T11:32:13.432324+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":15,"log_dups_size":0,"ondisk_log_size":15,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:56:19.989002+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":7,"num_object_clones":0,"num_object_copies":21,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":7,"num_whiteouts":0,"num_read":26,"num_read_kb":17,"num_write":16,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,2,0],"acting":[3,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.18","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058627+0000","last_change":"2026-03-10T11:32:09.514728+0000","last_active":"2026-03-10T11:32:14.058627+0000","last_peered":"2026-03-10T11:32:14.058627+0000","last_clean":"2026-03-10T11:32:14.058627+0000","last_became_active":"2026-03-10T11:32:09.509041+0000","last_became_peered":"2026-03-10T11:32:09.509041+0000","last_unstale":"2026-03-10T11:32:14.058627+0000","last_undegraded":"2026-03-10T11:32:14.058627+0000","last_fullsized":"2026-03-10T11:32:14.058627+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:23:50.009456+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,6,1],"acting":[4,6,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.1d","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434914+0000","last_change":"2026-03-10T11:32:05.477572+0000","last_active":"2026-03-10T11:32:13.434914+0000","last_peered":"2026-03-10T11:32:13.434914+0000","last_clean":"2026-03-10T11:32:13.434914+0000","last_became_active":"2026-03-10T11:32:05.477441+0000","last_became_peered":"2026-03-10T11:32:05.477441+0000","last_unstale":"2026-03-10T11:32:13.434914+0000","last_undegraded":"2026-03-10T11:32:13.434914+0000","last_fullsized":"2026-03-10T11:32:13.434914+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:49:49.474258+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,4,6],"acting":[5,4,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"4.1a","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058913+0000","last_change":"2026-03-10T11:32:07.480994+0000","last_active":"2026-03-10T11:32:14.058913+0000","last_peered":"2026-03-10T11:32:14.058913+0000","last_clean":"2026-03-10T11:32:14.058913+0000","last_became_active":"2026-03-10T11:32:07.480797+0000","last_became_peered":"2026-03-10T11:32:07.480797+0000","last_unstale":"2026-03-10T11:32:14.058913+0000","last_undegraded":"2026-03-10T11:32:14.058913+0000","last_fullsized":"2026-03-10T11:32:14.058913+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:08:46.328816+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,0],"acting":[4,3,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"5.1b","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434880+0000","last_change":"2026-03-10T11:32:09.513311+0000","last_active":"2026-03-10T11:32:13.434880+0000","last_peered":"2026-03-10T11:32:13.434880+0000","last_clean":"2026-03-10T11:32:13.434880+0000","last_became_active":"2026-03-10T11:32:09.513156+0000","last_became_peered":"2026-03-10T11:32:09.513156+0000","last_unstale":"2026-03-10T11:32:13.434880+0000","last_undegraded":"2026-03-10T11:32:13.434880+0000","last_fullsized":"2026-03-10T11:32:13.434880+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:07:26.592760+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,0,7],"acting":[5,0,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.18","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539342+0000","last_change":"2026-03-10T11:32:11.512406+0000","last_active":"2026-03-10T11:32:13.539342+0000","last_peered":"2026-03-10T11:32:13.539342+0000","last_clean":"2026-03-10T11:32:13.539342+0000","last_became_active":"2026-03-10T11:32:11.512324+0000","last_became_peered":"2026-03-10T11:32:11.512324+0000","last_unstale":"2026-03-10T11:32:13.539342+0000","last_undegraded":"2026-03-10T11:32:13.539342+0000","last_fullsized":"2026-03-10T11:32:13.539342+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:47:15.405374+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,1,7],"acting":[0,1,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"3.1c","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435186+0000","last_change":"2026-03-10T11:32:05.477512+0000","last_active":"2026-03-10T11:32:13.435186+0000","last_peered":"2026-03-10T11:32:13.435186+0000","last_clean":"2026-03-10T11:32:13.435186+0000","last_became_active":"2026-03-10T11:32:05.477337+0000","last_became_peered":"2026-03-10T11:32:05.477337+0000","last_unstale":"2026-03-10T11:32:13.435186+0000","last_undegraded":"2026-03-10T11:32:13.435186+0000","last_fullsized":"2026-03-10T11:32:13.435186+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:38:48.039126+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,4,1],"acting":[5,4,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"4.1b","version":"55'5","reported_seq":33,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058878+0000","last_change":"2026-03-10T11:32:07.483080+0000","last_active":"2026-03-10T11:32:14.058878+0000","last_peered":"2026-03-10T11:32:14.058878+0000","last_clean":"2026-03-10T11:32:14.058878+0000","last_became_active":"2026-03-10T11:32:07.482661+0000","last_became_peered":"2026-03-10T11:32:07.482661+0000","last_unstale":"2026-03-10T11:32:14.058878+0000","last_undegraded":"2026-03-10T11:32:14.058878+0000","last_fullsized":"2026-03-10T11:32:14.058878+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":5,"log_dups_size":0,"ondisk_log_size":5,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:56:09.413634+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":11,"num_read_kb":7,"num_write":6,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,1],"acting":[4,3,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"5.1a","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752686+0000","last_change":"2026-03-10T11:32:09.519196+0000","last_active":"2026-03-10T11:32:13.752686+0000","last_peered":"2026-03-10T11:32:13.752686+0000","last_clean":"2026-03-10T11:32:13.752686+0000","last_became_active":"2026-03-10T11:32:09.519015+0000","last_became_peered":"2026-03-10T11:32:09.519015+0000","last_unstale":"2026-03-10T11:32:13.752686+0000","last_undegraded":"2026-03-10T11:32:13.752686+0000","last_fullsized":"2026-03-10T11:32:13.752686+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:37:26.401819+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,1],"acting":[7,4,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"6.19","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435203+0000","last_change":"2026-03-10T11:32:11.505137+0000","last_active":"2026-03-10T11:32:13.435203+0000","last_peered":"2026-03-10T11:32:13.435203+0000","last_clean":"2026-03-10T11:32:13.435203+0000","last_became_active":"2026-03-10T11:32:11.504903+0000","last_became_peered":"2026-03-10T11:32:11.504903+0000","last_unstale":"2026-03-10T11:32:13.435203+0000","last_undegraded":"2026-03-10T11:32:13.435203+0000","last_fullsized":"2026-03-10T11:32:13.435203+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:14:18.453681+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1,3],"acting":[5,1,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.1e","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059775+0000","last_change":"2026-03-10T11:32:11.576043+0000","last_active":"2026-03-10T11:32:14.059775+0000","last_peered":"2026-03-10T11:32:14.059775+0000","last_clean":"2026-03-10T11:32:14.059775+0000","last_became_active":"2026-03-10T11:32:11.575896+0000","last_became_peered":"2026-03-10T11:32:11.575896+0000","last_unstale":"2026-03-10T11:32:14.059775+0000","last_undegraded":"2026-03-10T11:32:14.059775+0000","last_fullsized":"2026-03-10T11:32:14.059775+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:22:29.629653+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,6,5],"acting":[4,6,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.1b","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539184+0000","last_change":"2026-03-10T11:32:05.494175+0000","last_active":"2026-03-10T11:32:13.539184+0000","last_peered":"2026-03-10T11:32:13.539184+0000","last_clean":"2026-03-10T11:32:13.539184+0000","last_became_active":"2026-03-10T11:32:05.494030+0000","last_became_peered":"2026-03-10T11:32:05.494030+0000","last_unstale":"2026-03-10T11:32:13.539184+0000","last_undegraded":"2026-03-10T11:32:13.539184+0000","last_fullsized":"2026-03-10T11:32:13.539184+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:45:58.156722+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,4,7],"acting":[0,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.1c","version":"55'15","reported_seq":48,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095103+0000","last_change":"2026-03-10T11:32:07.483214+0000","last_active":"2026-03-10T11:32:14.095103+0000","last_peered":"2026-03-10T11:32:14.095103+0000","last_clean":"2026-03-10T11:32:14.095103+0000","last_became_active":"2026-03-10T11:32:07.483035+0000","last_became_peered":"2026-03-10T11:32:07.483035+0000","last_unstale":"2026-03-10T11:32:14.095103+0000","last_undegraded":"2026-03-10T11:32:14.095103+0000","last_fullsized":"2026-03-10T11:32:14.095103+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":15,"log_dups_size":0,"ondisk_log_size":15,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:40:44.403815+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":7,"num_object_clones":0,"num_object_copies":21,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":7,"num_whiteouts":0,"num_read":26,"num_read_kb":17,"num_write":16,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,1,3],"acting":[2,1,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"5.1d","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.756523+0000","last_change":"2026-03-10T11:32:09.519450+0000","last_active":"2026-03-10T11:32:13.756523+0000","last_peered":"2026-03-10T11:32:13.756523+0000","last_clean":"2026-03-10T11:32:13.756523+0000","last_became_active":"2026-03-10T11:32:09.518523+0000","last_became_peered":"2026-03-10T11:32:09.518523+0000","last_unstale":"2026-03-10T11:32:13.756523+0000","last_undegraded":"2026-03-10T11:32:13.756523+0000","last_fullsized":"2026-03-10T11:32:13.756523+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T23:29:05.618724+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,4,0],"acting":[1,4,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"6.1f","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.433035+0000","last_change":"2026-03-10T11:32:11.573792+0000","last_active":"2026-03-10T11:32:13.433035+0000","last_peered":"2026-03-10T11:32:13.433035+0000","last_clean":"2026-03-10T11:32:13.433035+0000","last_became_active":"2026-03-10T11:32:11.573629+0000","last_became_peered":"2026-03-10T11:32:11.573629+0000","last_unstale":"2026-03-10T11:32:13.433035+0000","last_undegraded":"2026-03-10T11:32:13.433035+0000","last_fullsized":"2026-03-10T11:32:13.433035+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:19:57.365774+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,6,5],"acting":[3,6,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.1a","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059058+0000","last_change":"2026-03-10T11:32:05.488341+0000","last_active":"2026-03-10T11:32:14.059058+0000","last_peered":"2026-03-10T11:32:14.059058+0000","last_clean":"2026-03-10T11:32:14.059058+0000","last_became_active":"2026-03-10T11:32:05.488227+0000","last_became_peered":"2026-03-10T11:32:05.488227+0000","last_unstale":"2026-03-10T11:32:14.059058+0000","last_undegraded":"2026-03-10T11:32:14.059058+0000","last_fullsized":"2026-03-10T11:32:14.059058+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:43:35.996147+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,1,2],"acting":[4,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"4.1d","version":"55'12","reported_seq":46,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.433064+0000","last_change":"2026-03-10T11:32:07.492198+0000","last_active":"2026-03-10T11:32:13.433064+0000","last_peered":"2026-03-10T11:32:13.433064+0000","last_clean":"2026-03-10T11:32:13.433064+0000","last_became_active":"2026-03-10T11:32:07.491637+0000","last_became_peered":"2026-03-10T11:32:07.491637+0000","last_unstale":"2026-03-10T11:32:13.433064+0000","last_undegraded":"2026-03-10T11:32:13.433064+0000","last_fullsized":"2026-03-10T11:32:13.433064+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":12,"log_dups_size":0,"ondisk_log_size":12,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:07:54.728750+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":220,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":25,"num_read_kb":16,"num_write":14,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,4],"acting":[3,1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.1c","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059038+0000","last_change":"2026-03-10T11:32:09.515333+0000","last_active":"2026-03-10T11:32:14.059038+0000","last_peered":"2026-03-10T11:32:14.059038+0000","last_clean":"2026-03-10T11:32:14.059038+0000","last_became_active":"2026-03-10T11:32:09.515091+0000","last_became_peered":"2026-03-10T11:32:09.515091+0000","last_unstale":"2026-03-10T11:32:14.059038+0000","last_undegraded":"2026-03-10T11:32:14.059038+0000","last_fullsized":"2026-03-10T11:32:14.059038+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:52:03.100394+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,2],"acting":[4,3,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"6.1c","version":"55'1","reported_seq":16,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753578+0000","last_change":"2026-03-10T11:32:11.514521+0000","last_active":"2026-03-10T11:32:13.753578+0000","last_peered":"2026-03-10T11:32:13.753578+0000","last_clean":"2026-03-10T11:32:13.753578+0000","last_became_active":"2026-03-10T11:32:11.513797+0000","last_became_peered":"2026-03-10T11:32:11.513797+0000","last_unstale":"2026-03-10T11:32:13.753578+0000","last_undegraded":"2026-03-10T11:32:13.753578+0000","last_fullsized":"2026-03-10T11:32:13.753578+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:55:33.816775+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":403,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,5,2],"acting":[7,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"3.19","version":"48'1","reported_seq":28,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.755975+0000","last_change":"2026-03-10T11:32:05.483058+0000","last_active":"2026-03-10T11:32:13.755975+0000","last_peered":"2026-03-10T11:32:13.755975+0000","last_clean":"2026-03-10T11:32:13.755975+0000","last_became_active":"2026-03-10T11:32:05.482922+0000","last_became_peered":"2026-03-10T11:32:05.482922+0000","last_unstale":"2026-03-10T11:32:13.755975+0000","last_undegraded":"2026-03-10T11:32:13.755975+0000","last_fullsized":"2026-03-10T11:32:13.755975+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:08:44.909272+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":46,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":1,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,3,4],"acting":[1,3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"4.1e","version":"55'10","reported_seq":38,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539696+0000","last_change":"2026-03-10T11:32:07.597568+0000","last_active":"2026-03-10T11:32:13.539696+0000","last_peered":"2026-03-10T11:32:13.539696+0000","last_clean":"2026-03-10T11:32:13.539696+0000","last_became_active":"2026-03-10T11:32:07.597489+0000","last_became_peered":"2026-03-10T11:32:07.597489+0000","last_unstale":"2026-03-10T11:32:13.539696+0000","last_undegraded":"2026-03-10T11:32:13.539696+0000","last_fullsized":"2026-03-10T11:32:13.539696+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":10,"log_dups_size":0,"ondisk_log_size":10,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:29:41.419571+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":15,"num_read_kb":10,"num_write":10,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,7,3],"acting":[0,7,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"5.1f","version":"55'8","reported_seq":33,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.430966+0000","last_change":"2026-03-10T11:32:09.517540+0000","last_active":"2026-03-10T11:32:13.430966+0000","last_peered":"2026-03-10T11:32:13.430966+0000","last_clean":"2026-03-10T11:32:13.430966+0000","last_became_active":"2026-03-10T11:32:09.517276+0000","last_became_peered":"2026-03-10T11:32:09.517276+0000","last_unstale":"2026-03-10T11:32:13.430966+0000","last_undegraded":"2026-03-10T11:32:13.430966+0000","last_fullsized":"2026-03-10T11:32:13.430966+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:58:11.682612+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,4,7],"acting":[6,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"4.f","version":"55'15","reported_seq":48,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.756635+0000","last_change":"2026-03-10T11:32:07.487724+0000","last_active":"2026-03-10T11:32:13.756635+0000","last_peered":"2026-03-10T11:32:13.756635+0000","last_clean":"2026-03-10T11:32:13.756635+0000","last_became_active":"2026-03-10T11:32:07.487306+0000","last_became_peered":"2026-03-10T11:32:07.487306+0000","last_unstale":"2026-03-10T11:32:13.756635+0000","last_undegraded":"2026-03-10T11:32:13.756635+0000","last_fullsized":"2026-03-10T11:32:13.756635+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":15,"log_dups_size":0,"ondisk_log_size":15,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:42:46.281406+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":7,"num_object_clones":0,"num_object_copies":21,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":7,"num_whiteouts":0,"num_read":26,"num_read_kb":17,"num_write":16,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,3,4],"acting":[1,3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.8","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432408+0000","last_change":"2026-03-10T11:32:05.492886+0000","last_active":"2026-03-10T11:32:13.432408+0000","last_peered":"2026-03-10T11:32:13.432408+0000","last_clean":"2026-03-10T11:32:13.432408+0000","last_became_active":"2026-03-10T11:32:05.492458+0000","last_became_peered":"2026-03-10T11:32:05.492458+0000","last_unstale":"2026-03-10T11:32:13.432408+0000","last_undegraded":"2026-03-10T11:32:13.432408+0000","last_fullsized":"2026-03-10T11:32:13.432408+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:01:32.990876+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,7],"acting":[3,1,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.e","version":"55'8","reported_seq":30,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058947+0000","last_change":"2026-03-10T11:32:09.515277+0000","last_active":"2026-03-10T11:32:14.058947+0000","last_peered":"2026-03-10T11:32:14.058947+0000","last_clean":"2026-03-10T11:32:14.058947+0000","last_became_active":"2026-03-10T11:32:09.514998+0000","last_became_peered":"2026-03-10T11:32:09.514998+0000","last_unstale":"2026-03-10T11:32:14.058947+0000","last_undegraded":"2026-03-10T11:32:14.058947+0000","last_fullsized":"2026-03-10T11:32:14.058947+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:43:50.105229+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,5,0],"acting":[4,5,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"6.d","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434790+0000","last_change":"2026-03-10T11:32:11.502209+0000","last_active":"2026-03-10T11:32:13.434790+0000","last_peered":"2026-03-10T11:32:13.434790+0000","last_clean":"2026-03-10T11:32:13.434790+0000","last_became_active":"2026-03-10T11:32:11.502106+0000","last_became_peered":"2026-03-10T11:32:11.502106+0000","last_unstale":"2026-03-10T11:32:13.434790+0000","last_undegraded":"2026-03-10T11:32:13.434790+0000","last_fullsized":"2026-03-10T11:32:13.434790+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:42:39.390458+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1,0],"acting":[5,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"4.0","version":"55'18","reported_seq":55,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432508+0000","last_change":"2026-03-10T11:32:07.598762+0000","last_active":"2026-03-10T11:32:13.432508+0000","last_peered":"2026-03-10T11:32:13.432508+0000","last_clean":"2026-03-10T11:32:13.432508+0000","last_became_active":"2026-03-10T11:32:07.598567+0000","last_became_peered":"2026-03-10T11:32:07.598567+0000","last_unstale":"2026-03-10T11:32:13.432508+0000","last_undegraded":"2026-03-10T11:32:13.432508+0000","last_fullsized":"2026-03-10T11:32:13.432508+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":18,"log_dups_size":0,"ondisk_log_size":18,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:45:00.149680+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":220,"num_objects":8,"num_object_clones":0,"num_object_copies":24,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":8,"num_whiteouts":0,"num_read":34,"num_read_kb":22,"num_write":20,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,0],"acting":[3,7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.7","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432476+0000","last_change":"2026-03-10T11:32:05.492977+0000","last_active":"2026-03-10T11:32:13.432476+0000","last_peered":"2026-03-10T11:32:13.432476+0000","last_clean":"2026-03-10T11:32:13.432476+0000","last_became_active":"2026-03-10T11:32:05.492594+0000","last_became_peered":"2026-03-10T11:32:05.492594+0000","last_unstale":"2026-03-10T11:32:13.432476+0000","last_undegraded":"2026-03-10T11:32:13.432476+0000","last_fullsized":"2026-03-10T11:32:13.432476+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:26:21.198593+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,0],"acting":[3,7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.1","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058622+0000","last_change":"2026-03-10T11:32:09.514857+0000","last_active":"2026-03-10T11:32:14.058622+0000","last_peered":"2026-03-10T11:32:14.058622+0000","last_clean":"2026-03-10T11:32:14.058622+0000","last_became_active":"2026-03-10T11:32:09.511003+0000","last_became_peered":"2026-03-10T11:32:09.511003+0000","last_unstale":"2026-03-10T11:32:14.058622+0000","last_undegraded":"2026-03-10T11:32:14.058622+0000","last_fullsized":"2026-03-10T11:32:14.058622+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:51:56.738928+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,7],"acting":[4,3,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"6.2","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058698+0000","last_change":"2026-03-10T11:32:11.510897+0000","last_active":"2026-03-10T11:32:14.058698+0000","last_peered":"2026-03-10T11:32:14.058698+0000","last_clean":"2026-03-10T11:32:14.058698+0000","last_became_active":"2026-03-10T11:32:11.510448+0000","last_became_peered":"2026-03-10T11:32:11.510448+0000","last_unstale":"2026-03-10T11:32:14.058698+0000","last_undegraded":"2026-03-10T11:32:14.058698+0000","last_fullsized":"2026-03-10T11:32:14.058698+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:35:04.941119+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,2],"acting":[4,3,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"4.1","version":"55'14","reported_seq":44,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059961+0000","last_change":"2026-03-10T11:32:07.479250+0000","last_active":"2026-03-10T11:32:14.059961+0000","last_peered":"2026-03-10T11:32:14.059961+0000","last_clean":"2026-03-10T11:32:14.059961+0000","last_became_active":"2026-03-10T11:32:07.478987+0000","last_became_peered":"2026-03-10T11:32:07.478987+0000","last_unstale":"2026-03-10T11:32:14.059961+0000","last_undegraded":"2026-03-10T11:32:14.059961+0000","last_fullsized":"2026-03-10T11:32:14.059961+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":14,"log_dups_size":0,"ondisk_log_size":14,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:14:05.132235+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":7,"num_object_clones":0,"num_object_copies":21,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":7,"num_whiteouts":0,"num_read":21,"num_read_kb":14,"num_write":14,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,5,6],"acting":[4,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.6","version":"48'1","reported_seq":28,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539218+0000","last_change":"2026-03-10T11:32:05.488767+0000","last_active":"2026-03-10T11:32:13.539218+0000","last_peered":"2026-03-10T11:32:13.539218+0000","last_clean":"2026-03-10T11:32:13.539218+0000","last_became_active":"2026-03-10T11:32:05.488670+0000","last_became_peered":"2026-03-10T11:32:05.488670+0000","last_unstale":"2026-03-10T11:32:13.539218+0000","last_undegraded":"2026-03-10T11:32:13.539218+0000","last_fullsized":"2026-03-10T11:32:13.539218+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:38:01.875008+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":46,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":1,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,1,4],"acting":[0,1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"5.0","version":"55'8","reported_seq":30,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431993+0000","last_change":"2026-03-10T11:32:09.495267+0000","last_active":"2026-03-10T11:32:13.431993+0000","last_peered":"2026-03-10T11:32:13.431993+0000","last_clean":"2026-03-10T11:32:13.431993+0000","last_became_active":"2026-03-10T11:32:09.495115+0000","last_became_peered":"2026-03-10T11:32:09.495115+0000","last_unstale":"2026-03-10T11:32:13.431993+0000","last_undegraded":"2026-03-10T11:32:13.431993+0000","last_fullsized":"2026-03-10T11:32:13.431993+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:52:02.237765+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,4],"acting":[3,1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"6.3","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752423+0000","last_change":"2026-03-10T11:32:11.578302+0000","last_active":"2026-03-10T11:32:13.752423+0000","last_peered":"2026-03-10T11:32:13.752423+0000","last_clean":"2026-03-10T11:32:13.752423+0000","last_became_active":"2026-03-10T11:32:11.577452+0000","last_became_peered":"2026-03-10T11:32:11.577452+0000","last_unstale":"2026-03-10T11:32:13.752423+0000","last_undegraded":"2026-03-10T11:32:13.752423+0000","last_fullsized":"2026-03-10T11:32:13.752423+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:24:57.973295+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,6,2],"acting":[7,6,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"4.2","version":"55'10","reported_seq":38,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757315+0000","last_change":"2026-03-10T11:32:07.488057+0000","last_active":"2026-03-10T11:32:13.757315+0000","last_peered":"2026-03-10T11:32:13.757315+0000","last_clean":"2026-03-10T11:32:13.757315+0000","last_became_active":"2026-03-10T11:32:07.487903+0000","last_became_peered":"2026-03-10T11:32:07.487903+0000","last_unstale":"2026-03-10T11:32:13.757315+0000","last_undegraded":"2026-03-10T11:32:13.757315+0000","last_fullsized":"2026-03-10T11:32:13.757315+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":10,"log_dups_size":0,"ondisk_log_size":10,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:22:35.288282+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":15,"num_read_kb":10,"num_write":10,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,4],"acting":[1,5,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.5","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435492+0000","last_change":"2026-03-10T11:32:05.486010+0000","last_active":"2026-03-10T11:32:13.435492+0000","last_peered":"2026-03-10T11:32:13.435492+0000","last_clean":"2026-03-10T11:32:13.435492+0000","last_became_active":"2026-03-10T11:32:05.485734+0000","last_became_peered":"2026-03-10T11:32:05.485734+0000","last_unstale":"2026-03-10T11:32:13.435492+0000","last_undegraded":"2026-03-10T11:32:13.435492+0000","last_fullsized":"2026-03-10T11:32:13.435492+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:58:01.385548+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,3,2],"acting":[5,3,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"5.3","version":"55'8","reported_seq":30,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539874+0000","last_change":"2026-03-10T11:32:09.505558+0000","last_active":"2026-03-10T11:32:13.539874+0000","last_peered":"2026-03-10T11:32:13.539874+0000","last_clean":"2026-03-10T11:32:13.539874+0000","last_became_active":"2026-03-10T11:32:09.505467+0000","last_became_peered":"2026-03-10T11:32:09.505467+0000","last_unstale":"2026-03-10T11:32:13.539874+0000","last_undegraded":"2026-03-10T11:32:13.539874+0000","last_fullsized":"2026-03-10T11:32:13.539874+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:27:40.165203+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,6,5],"acting":[0,6,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"6.0","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539722+0000","last_change":"2026-03-10T11:32:11.524869+0000","last_active":"2026-03-10T11:32:13.539722+0000","last_peered":"2026-03-10T11:32:13.539722+0000","last_clean":"2026-03-10T11:32:13.539722+0000","last_became_active":"2026-03-10T11:32:11.524764+0000","last_became_peered":"2026-03-10T11:32:11.524764+0000","last_unstale":"2026-03-10T11:32:13.539722+0000","last_undegraded":"2026-03-10T11:32:13.539722+0000","last_fullsized":"2026-03-10T11:32:13.539722+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:50:58.880334+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,3,2],"acting":[0,3,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.3","version":"55'19","reported_seq":59,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539585+0000","last_change":"2026-03-10T11:32:07.597361+0000","last_active":"2026-03-10T11:32:13.539585+0000","last_peered":"2026-03-10T11:32:13.539585+0000","last_clean":"2026-03-10T11:32:13.539585+0000","last_became_active":"2026-03-10T11:32:07.597199+0000","last_became_peered":"2026-03-10T11:32:07.597199+0000","last_unstale":"2026-03-10T11:32:13.539585+0000","last_undegraded":"2026-03-10T11:32:13.539585+0000","last_fullsized":"2026-03-10T11:32:13.539585+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":19,"log_dups_size":0,"ondisk_log_size":19,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:31:09.996144+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":330,"num_objects":8,"num_object_clones":0,"num_object_copies":24,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":8,"num_whiteouts":0,"num_read":39,"num_read_kb":25,"num_write":22,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,5,7],"acting":[0,5,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"3.4","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.755928+0000","last_change":"2026-03-10T11:32:05.470777+0000","last_active":"2026-03-10T11:32:13.755928+0000","last_peered":"2026-03-10T11:32:13.755928+0000","last_clean":"2026-03-10T11:32:13.755928+0000","last_became_active":"2026-03-10T11:32:05.469871+0000","last_became_peered":"2026-03-10T11:32:05.469871+0000","last_unstale":"2026-03-10T11:32:13.755928+0000","last_undegraded":"2026-03-10T11:32:13.755928+0000","last_fullsized":"2026-03-10T11:32:13.755928+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:40:37.152961+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,5],"acting":[1,2,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"5.2","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431003+0000","last_change":"2026-03-10T11:32:09.517469+0000","last_active":"2026-03-10T11:32:13.431003+0000","last_peered":"2026-03-10T11:32:13.431003+0000","last_clean":"2026-03-10T11:32:13.431003+0000","last_became_active":"2026-03-10T11:32:09.517068+0000","last_became_peered":"2026-03-10T11:32:09.517068+0000","last_unstale":"2026-03-10T11:32:13.431003+0000","last_undegraded":"2026-03-10T11:32:13.431003+0000","last_fullsized":"2026-03-10T11:32:13.431003+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:59:19.119041+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,0,5],"acting":[6,0,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"6.1","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757256+0000","last_change":"2026-03-10T11:32:11.575149+0000","last_active":"2026-03-10T11:32:13.757256+0000","last_peered":"2026-03-10T11:32:13.757256+0000","last_clean":"2026-03-10T11:32:13.757256+0000","last_became_active":"2026-03-10T11:32:11.574450+0000","last_became_peered":"2026-03-10T11:32:11.574450+0000","last_unstale":"2026-03-10T11:32:13.757256+0000","last_undegraded":"2026-03-10T11:32:13.757256+0000","last_fullsized":"2026-03-10T11:32:13.757256+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:44:00.507675+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,6,2],"acting":[1,6,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"4.4","version":"55'28","reported_seq":74,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757202+0000","last_change":"2026-03-10T11:32:07.486298+0000","last_active":"2026-03-10T11:32:13.757202+0000","last_peered":"2026-03-10T11:32:13.757202+0000","last_clean":"2026-03-10T11:32:13.757202+0000","last_became_active":"2026-03-10T11:32:07.486212+0000","last_became_peered":"2026-03-10T11:32:07.486212+0000","last_unstale":"2026-03-10T11:32:13.757202+0000","last_undegraded":"2026-03-10T11:32:13.757202+0000","last_fullsized":"2026-03-10T11:32:13.757202+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":28,"log_dups_size":0,"ondisk_log_size":28,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:37:25.312706+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":358,"num_objects":10,"num_object_clones":0,"num_object_copies":30,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":10,"num_whiteouts":0,"num_read":48,"num_read_kb":33,"num_write":26,"num_write_kb":4,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,3],"acting":[1,2,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.3","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058290+0000","last_change":"2026-03-10T11:32:05.482284+0000","last_active":"2026-03-10T11:32:14.058290+0000","last_peered":"2026-03-10T11:32:14.058290+0000","last_clean":"2026-03-10T11:32:14.058290+0000","last_became_active":"2026-03-10T11:32:05.482174+0000","last_became_peered":"2026-03-10T11:32:05.482174+0000","last_unstale":"2026-03-10T11:32:14.058290+0000","last_undegraded":"2026-03-10T11:32:14.058290+0000","last_fullsized":"2026-03-10T11:32:14.058290+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:28:56.826046+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,0,6],"acting":[4,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"2.2","version":"50'2","reported_seq":34,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435583+0000","last_change":"2026-03-10T11:32:07.468273+0000","last_active":"2026-03-10T11:32:13.435583+0000","last_peered":"2026-03-10T11:32:13.435583+0000","last_clean":"2026-03-10T11:32:13.435583+0000","last_became_active":"2026-03-10T11:32:05.466762+0000","last_became_peered":"2026-03-10T11:32:05.466762+0000","last_unstale":"2026-03-10T11:32:13.435583+0000","last_undegraded":"2026-03-10T11:32:13.435583+0000","last_fullsized":"2026-03-10T11:32:13.435583+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:51:31.016672+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00040810500000000002,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1,6],"acting":[5,1,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"5.5","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539487+0000","last_change":"2026-03-10T11:32:09.515802+0000","last_active":"2026-03-10T11:32:13.539487+0000","last_peered":"2026-03-10T11:32:13.539487+0000","last_clean":"2026-03-10T11:32:13.539487+0000","last_became_active":"2026-03-10T11:32:09.515691+0000","last_became_peered":"2026-03-10T11:32:09.515691+0000","last_unstale":"2026-03-10T11:32:13.539487+0000","last_undegraded":"2026-03-10T11:32:13.539487+0000","last_fullsized":"2026-03-10T11:32:13.539487+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:34:27.429654+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,1,4],"acting":[0,1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"6.6","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432064+0000","last_change":"2026-03-10T11:32:11.509752+0000","last_active":"2026-03-10T11:32:13.432064+0000","last_peered":"2026-03-10T11:32:13.432064+0000","last_clean":"2026-03-10T11:32:13.432064+0000","last_became_active":"2026-03-10T11:32:11.509288+0000","last_became_peered":"2026-03-10T11:32:11.509288+0000","last_unstale":"2026-03-10T11:32:13.432064+0000","last_undegraded":"2026-03-10T11:32:13.432064+0000","last_fullsized":"2026-03-10T11:32:13.432064+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:25:33.241815+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,4,7],"acting":[3,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.7","version":"55'13","reported_seq":50,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757287+0000","last_change":"2026-03-10T11:32:07.485025+0000","last_active":"2026-03-10T11:32:13.757287+0000","last_peered":"2026-03-10T11:32:13.757287+0000","last_clean":"2026-03-10T11:32:13.757287+0000","last_became_active":"2026-03-10T11:32:07.484944+0000","last_became_peered":"2026-03-10T11:32:07.484944+0000","last_unstale":"2026-03-10T11:32:13.757287+0000","last_undegraded":"2026-03-10T11:32:13.757287+0000","last_fullsized":"2026-03-10T11:32:13.757287+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":13,"log_dups_size":0,"ondisk_log_size":13,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:49:17.507119+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":330,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":30,"num_read_kb":19,"num_write":16,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,0],"acting":[1,5,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.0","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.755868+0000","last_change":"2026-03-10T11:32:05.469815+0000","last_active":"2026-03-10T11:32:13.755868+0000","last_peered":"2026-03-10T11:32:13.755868+0000","last_clean":"2026-03-10T11:32:13.755868+0000","last_became_active":"2026-03-10T11:32:05.469717+0000","last_became_peered":"2026-03-10T11:32:05.469717+0000","last_unstale":"2026-03-10T11:32:13.755868+0000","last_undegraded":"2026-03-10T11:32:13.755868+0000","last_fullsized":"2026-03-10T11:32:13.755868+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:23:39.249935+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,6],"acting":[1,2,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"2.1","version":"48'1","reported_seq":33,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.094602+0000","last_change":"2026-03-10T11:32:07.469162+0000","last_active":"2026-03-10T11:32:14.094602+0000","last_peered":"2026-03-10T11:32:14.094602+0000","last_clean":"2026-03-10T11:32:14.094602+0000","last_became_active":"2026-03-10T11:32:05.481644+0000","last_became_peered":"2026-03-10T11:32:05.481644+0000","last_unstale":"2026-03-10T11:32:14.094602+0000","last_undegraded":"2026-03-10T11:32:14.094602+0000","last_fullsized":"2026-03-10T11:32:14.094602+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:54:36.058555+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00034258999999999998,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,3,0],"acting":[2,3,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"5.6","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095075+0000","last_change":"2026-03-10T11:32:09.504060+0000","last_active":"2026-03-10T11:32:14.095075+0000","last_peered":"2026-03-10T11:32:14.095075+0000","last_clean":"2026-03-10T11:32:14.095075+0000","last_became_active":"2026-03-10T11:32:09.503889+0000","last_became_peered":"2026-03-10T11:32:09.503889+0000","last_unstale":"2026-03-10T11:32:14.095075+0000","last_undegraded":"2026-03-10T11:32:14.095075+0000","last_fullsized":"2026-03-10T11:32:14.095075+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:29:57.323993+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,5,7],"acting":[2,5,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"6.5","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752569+0000","last_change":"2026-03-10T11:32:11.577427+0000","last_active":"2026-03-10T11:32:13.752569+0000","last_peered":"2026-03-10T11:32:13.752569+0000","last_clean":"2026-03-10T11:32:13.752569+0000","last_became_active":"2026-03-10T11:32:11.576376+0000","last_became_peered":"2026-03-10T11:32:11.576376+0000","last_unstale":"2026-03-10T11:32:13.752569+0000","last_undegraded":"2026-03-10T11:32:13.752569+0000","last_fullsized":"2026-03-10T11:32:13.752569+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:12:59.201772+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,6,3],"acting":[7,6,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"4.6","version":"55'12","reported_seq":41,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539292+0000","last_change":"2026-03-10T11:32:07.491133+0000","last_active":"2026-03-10T11:32:13.539292+0000","last_peered":"2026-03-10T11:32:13.539292+0000","last_clean":"2026-03-10T11:32:13.539292+0000","last_became_active":"2026-03-10T11:32:07.490960+0000","last_became_peered":"2026-03-10T11:32:07.490960+0000","last_unstale":"2026-03-10T11:32:13.539292+0000","last_undegraded":"2026-03-10T11:32:13.539292+0000","last_fullsized":"2026-03-10T11:32:13.539292+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":12,"log_dups_size":0,"ondisk_log_size":12,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:52:08.881689+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":6,"num_object_clones":0,"num_object_copies":18,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":6,"num_whiteouts":0,"num_read":18,"num_read_kb":12,"num_write":12,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,1,2],"acting":[0,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"3.1","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.538897+0000","last_change":"2026-03-10T11:32:05.484386+0000","last_active":"2026-03-10T11:32:13.538897+0000","last_peered":"2026-03-10T11:32:13.538897+0000","last_clean":"2026-03-10T11:32:13.538897+0000","last_became_active":"2026-03-10T11:32:05.484276+0000","last_became_peered":"2026-03-10T11:32:05.484276+0000","last_unstale":"2026-03-10T11:32:13.538897+0000","last_undegraded":"2026-03-10T11:32:13.538897+0000","last_fullsized":"2026-03-10T11:32:13.538897+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T23:27:09.503732+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,4,3],"acting":[0,4,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"2.0","version":"56'5","reported_seq":43,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:15.198196+0000","last_change":"2026-03-10T11:32:07.595732+0000","last_active":"2026-03-10T11:32:15.198196+0000","last_peered":"2026-03-10T11:32:15.198196+0000","last_clean":"2026-03-10T11:32:15.198196+0000","last_became_active":"2026-03-10T11:32:05.484114+0000","last_became_peered":"2026-03-10T11:32:05.484114+0000","last_unstale":"2026-03-10T11:32:15.198196+0000","last_undegraded":"2026-03-10T11:32:15.198196+0000","last_fullsized":"2026-03-10T11:32:15.198196+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":5,"log_dups_size":0,"ondisk_log_size":5,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:21:40.405562+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00025977500000000001,"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":8,"num_read_kb":3,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,1,0],"acting":[7,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.7","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435296+0000","last_change":"2026-03-10T11:32:09.500766+0000","last_active":"2026-03-10T11:32:13.435296+0000","last_peered":"2026-03-10T11:32:13.435296+0000","last_clean":"2026-03-10T11:32:13.435296+0000","last_became_active":"2026-03-10T11:32:09.500135+0000","last_became_peered":"2026-03-10T11:32:09.500135+0000","last_unstale":"2026-03-10T11:32:13.435296+0000","last_undegraded":"2026-03-10T11:32:13.435296+0000","last_fullsized":"2026-03-10T11:32:13.435296+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:02:55.101840+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1,0],"acting":[5,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.4","version":"55'1","reported_seq":16,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.756803+0000","last_change":"2026-03-10T11:32:11.515730+0000","last_active":"2026-03-10T11:32:13.756803+0000","last_peered":"2026-03-10T11:32:13.756803+0000","last_clean":"2026-03-10T11:32:13.756803+0000","last_became_active":"2026-03-10T11:32:11.511711+0000","last_became_peered":"2026-03-10T11:32:11.511711+0000","last_unstale":"2026-03-10T11:32:13.756803+0000","last_undegraded":"2026-03-10T11:32:13.756803+0000","last_fullsized":"2026-03-10T11:32:13.756803+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:39:45.795505+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":13,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":1,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,3],"acting":[1,5,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"4.5","version":"55'16","reported_seq":48,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431268+0000","last_change":"2026-03-10T11:32:07.597624+0000","last_active":"2026-03-10T11:32:13.431268+0000","last_peered":"2026-03-10T11:32:13.431268+0000","last_clean":"2026-03-10T11:32:13.431268+0000","last_became_active":"2026-03-10T11:32:07.597049+0000","last_became_peered":"2026-03-10T11:32:07.597049+0000","last_unstale":"2026-03-10T11:32:13.431268+0000","last_undegraded":"2026-03-10T11:32:13.431268+0000","last_fullsized":"2026-03-10T11:32:13.431268+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":16,"log_dups_size":0,"ondisk_log_size":16,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:17:30.419013+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":154,"num_objects":8,"num_object_clones":0,"num_object_copies":24,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":8,"num_whiteouts":0,"num_read":25,"num_read_kb":15,"num_write":13,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,4,7],"acting":[6,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"3.2","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432591+0000","last_change":"2026-03-10T11:32:05.483910+0000","last_active":"2026-03-10T11:32:13.432591+0000","last_peered":"2026-03-10T11:32:13.432591+0000","last_clean":"2026-03-10T11:32:13.432591+0000","last_became_active":"2026-03-10T11:32:05.480854+0000","last_became_peered":"2026-03-10T11:32:05.480854+0000","last_unstale":"2026-03-10T11:32:13.432591+0000","last_undegraded":"2026-03-10T11:32:13.432591+0000","last_fullsized":"2026-03-10T11:32:13.432591+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:44:44.275870+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,6],"acting":[3,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"1.0","version":"18'32","reported_seq":37,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753175+0000","last_change":"2026-03-10T11:32:03.457270+0000","last_active":"2026-03-10T11:32:13.753175+0000","last_peered":"2026-03-10T11:32:13.753175+0000","last_clean":"2026-03-10T11:32:13.753175+0000","last_became_active":"2026-03-10T11:32:03.451510+0000","last_became_peered":"2026-03-10T11:32:03.451510+0000","last_unstale":"2026-03-10T11:32:13.753175+0000","last_undegraded":"2026-03-10T11:32:13.753175+0000","last_fullsized":"2026-03-10T11:32:13.753175+0000","mapping_epoch":45,"log_start":"0'0","ondisk_log_start":"0'0","created":17,"last_epoch_clean":46,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:31:09.189169+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:31:09.189169+0000","last_clean_scrub_stamp":"2026-03-10T11:31:09.189169+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:34:21.433105+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":459280,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0,6],"acting":[7,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.4","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753235+0000","last_change":"2026-03-10T11:32:09.525158+0000","last_active":"2026-03-10T11:32:13.753235+0000","last_peered":"2026-03-10T11:32:13.753235+0000","last_clean":"2026-03-10T11:32:13.753235+0000","last_became_active":"2026-03-10T11:32:09.525051+0000","last_became_peered":"2026-03-10T11:32:09.525051+0000","last_unstale":"2026-03-10T11:32:13.753235+0000","last_undegraded":"2026-03-10T11:32:13.753235+0000","last_fullsized":"2026-03-10T11:32:13.753235+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:51:00.114489+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,2,5],"acting":[7,2,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"6.7","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434911+0000","last_change":"2026-03-10T11:32:11.505220+0000","last_active":"2026-03-10T11:32:13.434911+0000","last_peered":"2026-03-10T11:32:13.434911+0000","last_clean":"2026-03-10T11:32:13.434911+0000","last_became_active":"2026-03-10T11:32:11.505041+0000","last_became_peered":"2026-03-10T11:32:11.505041+0000","last_unstale":"2026-03-10T11:32:13.434911+0000","last_undegraded":"2026-03-10T11:32:13.434911+0000","last_fullsized":"2026-03-10T11:32:13.434911+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:20:43.846661+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,3,4],"acting":[5,3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"4.e","version":"55'11","reported_seq":42,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059204+0000","last_change":"2026-03-10T11:32:07.482751+0000","last_active":"2026-03-10T11:32:14.059204+0000","last_peered":"2026-03-10T11:32:14.059204+0000","last_clean":"2026-03-10T11:32:14.059204+0000","last_became_active":"2026-03-10T11:32:07.482298+0000","last_became_peered":"2026-03-10T11:32:07.482298+0000","last_unstale":"2026-03-10T11:32:14.059204+0000","last_undegraded":"2026-03-10T11:32:14.059204+0000","last_fullsized":"2026-03-10T11:32:14.059204+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":11,"log_dups_size":0,"ondisk_log_size":11,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:29:34.576488+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":20,"num_read_kb":13,"num_write":12,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,6,1],"acting":[4,6,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.9","version":"48'1","reported_seq":33,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059171+0000","last_change":"2026-03-10T11:32:05.492929+0000","last_active":"2026-03-10T11:32:14.059171+0000","last_peered":"2026-03-10T11:32:14.059171+0000","last_clean":"2026-03-10T11:32:14.059171+0000","last_became_active":"2026-03-10T11:32:05.492820+0000","last_became_peered":"2026-03-10T11:32:05.492820+0000","last_unstale":"2026-03-10T11:32:14.059171+0000","last_undegraded":"2026-03-10T11:32:14.059171+0000","last_fullsized":"2026-03-10T11:32:14.059171+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:00:17.535577+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":993,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":7,"num_read_kb":7,"num_write":2,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,2,7],"acting":[4,2,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"5.f","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435749+0000","last_change":"2026-03-10T11:32:09.513230+0000","last_active":"2026-03-10T11:32:13.435749+0000","last_peered":"2026-03-10T11:32:13.435749+0000","last_clean":"2026-03-10T11:32:13.435749+0000","last_became_active":"2026-03-10T11:32:09.513045+0000","last_became_peered":"2026-03-10T11:32:09.513045+0000","last_unstale":"2026-03-10T11:32:13.435749+0000","last_undegraded":"2026-03-10T11:32:13.435749+0000","last_fullsized":"2026-03-10T11:32:13.435749+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:48:47.617831+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,4,6],"acting":[5,4,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.c","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431921+0000","last_change":"2026-03-10T11:32:11.573142+0000","last_active":"2026-03-10T11:32:13.431921+0000","last_peered":"2026-03-10T11:32:13.431921+0000","last_clean":"2026-03-10T11:32:13.431921+0000","last_became_active":"2026-03-10T11:32:11.573026+0000","last_became_peered":"2026-03-10T11:32:11.573026+0000","last_unstale":"2026-03-10T11:32:13.431921+0000","last_undegraded":"2026-03-10T11:32:13.431921+0000","last_fullsized":"2026-03-10T11:32:13.431921+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:18:04.964512+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,6,5],"acting":[3,6,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.d","version":"55'17","reported_seq":51,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059505+0000","last_change":"2026-03-10T11:32:07.482964+0000","last_active":"2026-03-10T11:32:14.059505+0000","last_peered":"2026-03-10T11:32:14.059505+0000","last_clean":"2026-03-10T11:32:14.059505+0000","last_became_active":"2026-03-10T11:32:07.482540+0000","last_became_peered":"2026-03-10T11:32:07.482540+0000","last_unstale":"2026-03-10T11:32:14.059505+0000","last_undegraded":"2026-03-10T11:32:14.059505+0000","last_fullsized":"2026-03-10T11:32:14.059505+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":17,"log_dups_size":0,"ondisk_log_size":17,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:51:57.573737+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":8,"num_object_clones":0,"num_object_copies":24,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":8,"num_whiteouts":0,"num_read":29,"num_read_kb":19,"num_write":18,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,2,1],"acting":[4,2,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.a","version":"48'1","reported_seq":33,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431472+0000","last_change":"2026-03-10T11:32:05.482905+0000","last_active":"2026-03-10T11:32:13.431472+0000","last_peered":"2026-03-10T11:32:13.431472+0000","last_clean":"2026-03-10T11:32:13.431472+0000","last_became_active":"2026-03-10T11:32:05.482830+0000","last_became_peered":"2026-03-10T11:32:05.482830+0000","last_unstale":"2026-03-10T11:32:13.431472+0000","last_undegraded":"2026-03-10T11:32:13.431472+0000","last_fullsized":"2026-03-10T11:32:13.431472+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:41:28.409216+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":436,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":7,"num_read_kb":7,"num_write":2,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,4,1],"acting":[6,4,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"5.c","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.756856+0000","last_change":"2026-03-10T11:32:09.518756+0000","last_active":"2026-03-10T11:32:13.756856+0000","last_peered":"2026-03-10T11:32:13.756856+0000","last_clean":"2026-03-10T11:32:13.756856+0000","last_became_active":"2026-03-10T11:32:09.518665+0000","last_became_peered":"2026-03-10T11:32:09.518665+0000","last_unstale":"2026-03-10T11:32:13.756856+0000","last_undegraded":"2026-03-10T11:32:13.756856+0000","last_fullsized":"2026-03-10T11:32:13.756856+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:18:11.441776+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,4,0],"acting":[1,4,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"6.f","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095132+0000","last_change":"2026-03-10T11:32:11.527689+0000","last_active":"2026-03-10T11:32:14.095132+0000","last_peered":"2026-03-10T11:32:14.095132+0000","last_clean":"2026-03-10T11:32:14.095132+0000","last_became_active":"2026-03-10T11:32:11.527559+0000","last_became_peered":"2026-03-10T11:32:11.527559+0000","last_unstale":"2026-03-10T11:32:14.095132+0000","last_undegraded":"2026-03-10T11:32:14.095132+0000","last_fullsized":"2026-03-10T11:32:14.095132+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:29:29.018133+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,3,4],"acting":[2,3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"4.c","version":"55'10","reported_seq":38,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059742+0000","last_change":"2026-03-10T11:32:07.481053+0000","last_active":"2026-03-10T11:32:14.059742+0000","last_peered":"2026-03-10T11:32:14.059742+0000","last_clean":"2026-03-10T11:32:14.059742+0000","last_became_active":"2026-03-10T11:32:07.480920+0000","last_became_peered":"2026-03-10T11:32:07.480920+0000","last_unstale":"2026-03-10T11:32:14.059742+0000","last_undegraded":"2026-03-10T11:32:14.059742+0000","last_fullsized":"2026-03-10T11:32:14.059742+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":10,"log_dups_size":0,"ondisk_log_size":10,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:44:15.363985+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":15,"num_read_kb":10,"num_write":10,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,3,6],"acting":[4,3,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.b","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432351+0000","last_change":"2026-03-10T11:32:05.484342+0000","last_active":"2026-03-10T11:32:13.432351+0000","last_peered":"2026-03-10T11:32:13.432351+0000","last_clean":"2026-03-10T11:32:13.432351+0000","last_became_active":"2026-03-10T11:32:05.484242+0000","last_became_peered":"2026-03-10T11:32:05.484242+0000","last_unstale":"2026-03-10T11:32:13.432351+0000","last_undegraded":"2026-03-10T11:32:13.432351+0000","last_fullsized":"2026-03-10T11:32:13.432351+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:32:44.695262+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0,4],"acting":[3,0,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.d","version":"55'8","reported_seq":30,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095358+0000","last_change":"2026-03-10T11:32:09.504159+0000","last_active":"2026-03-10T11:32:14.095358+0000","last_peered":"2026-03-10T11:32:14.095358+0000","last_clean":"2026-03-10T11:32:14.095358+0000","last_became_active":"2026-03-10T11:32:09.503998+0000","last_became_peered":"2026-03-10T11:32:09.503998+0000","last_unstale":"2026-03-10T11:32:14.095358+0000","last_undegraded":"2026-03-10T11:32:14.095358+0000","last_fullsized":"2026-03-10T11:32:14.095358+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:09:16.830610+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,7,5],"acting":[2,7,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"6.e","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059642+0000","last_change":"2026-03-10T11:32:11.512156+0000","last_active":"2026-03-10T11:32:14.059642+0000","last_peered":"2026-03-10T11:32:14.059642+0000","last_clean":"2026-03-10T11:32:14.059642+0000","last_became_active":"2026-03-10T11:32:11.512023+0000","last_became_peered":"2026-03-10T11:32:11.512023+0000","last_unstale":"2026-03-10T11:32:14.059642+0000","last_undegraded":"2026-03-10T11:32:14.059642+0000","last_fullsized":"2026-03-10T11:32:14.059642+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:31:40.808175+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,1,2],"acting":[4,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"4.b","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539640+0000","last_change":"2026-03-10T11:32:07.490958+0000","last_active":"2026-03-10T11:32:13.539640+0000","last_peered":"2026-03-10T11:32:13.539640+0000","last_clean":"2026-03-10T11:32:13.539640+0000","last_became_active":"2026-03-10T11:32:07.490846+0000","last_became_peered":"2026-03-10T11:32:07.490846+0000","last_unstale":"2026-03-10T11:32:13.539640+0000","last_undegraded":"2026-03-10T11:32:13.539640+0000","last_fullsized":"2026-03-10T11:32:13.539640+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:02:19.295560+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,1,4],"acting":[0,1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"3.c","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435404+0000","last_change":"2026-03-10T11:32:05.486072+0000","last_active":"2026-03-10T11:32:13.435404+0000","last_peered":"2026-03-10T11:32:13.435404+0000","last_clean":"2026-03-10T11:32:13.435404+0000","last_became_active":"2026-03-10T11:32:05.485842+0000","last_became_peered":"2026-03-10T11:32:05.485842+0000","last_unstale":"2026-03-10T11:32:13.435404+0000","last_undegraded":"2026-03-10T11:32:13.435404+0000","last_fullsized":"2026-03-10T11:32:13.435404+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:34:56.952419+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,3,6],"acting":[5,3,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"5.a","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095262+0000","last_change":"2026-03-10T11:32:09.492441+0000","last_active":"2026-03-10T11:32:14.095262+0000","last_peered":"2026-03-10T11:32:14.095262+0000","last_clean":"2026-03-10T11:32:14.095262+0000","last_became_active":"2026-03-10T11:32:09.492151+0000","last_became_peered":"2026-03-10T11:32:09.492151+0000","last_unstale":"2026-03-10T11:32:14.095262+0000","last_undegraded":"2026-03-10T11:32:14.095262+0000","last_fullsized":"2026-03-10T11:32:14.095262+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:16:19.045768+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,4,3],"acting":[2,4,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"6.9","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539613+0000","last_change":"2026-03-10T11:32:11.502345+0000","last_active":"2026-03-10T11:32:13.539613+0000","last_peered":"2026-03-10T11:32:13.539613+0000","last_clean":"2026-03-10T11:32:13.539613+0000","last_became_active":"2026-03-10T11:32:11.502116+0000","last_became_peered":"2026-03-10T11:32:11.502116+0000","last_unstale":"2026-03-10T11:32:13.539613+0000","last_undegraded":"2026-03-10T11:32:13.539613+0000","last_fullsized":"2026-03-10T11:32:13.539613+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:16:04.120385+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,7,2],"acting":[0,7,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.a","version":"55'19","reported_seq":54,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431517+0000","last_change":"2026-03-10T11:32:07.597136+0000","last_active":"2026-03-10T11:32:13.431517+0000","last_peered":"2026-03-10T11:32:13.431517+0000","last_clean":"2026-03-10T11:32:13.431517+0000","last_became_active":"2026-03-10T11:32:07.596902+0000","last_became_peered":"2026-03-10T11:32:07.596902+0000","last_unstale":"2026-03-10T11:32:13.431517+0000","last_undegraded":"2026-03-10T11:32:13.431517+0000","last_fullsized":"2026-03-10T11:32:13.431517+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":19,"log_dups_size":0,"ondisk_log_size":19,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:06:20.801200+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":9,"num_object_clones":0,"num_object_copies":27,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":9,"num_whiteouts":0,"num_read":32,"num_read_kb":21,"num_write":20,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,1,7],"acting":[6,1,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"3.d","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752240+0000","last_change":"2026-03-10T11:32:05.484420+0000","last_active":"2026-03-10T11:32:13.752240+0000","last_peered":"2026-03-10T11:32:13.752240+0000","last_clean":"2026-03-10T11:32:13.752240+0000","last_became_active":"2026-03-10T11:32:05.484301+0000","last_became_peered":"2026-03-10T11:32:05.484301+0000","last_unstale":"2026-03-10T11:32:13.752240+0000","last_undegraded":"2026-03-10T11:32:13.752240+0000","last_fullsized":"2026-03-10T11:32:13.752240+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:56:27.516701+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,5,6],"acting":[7,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.b","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095302+0000","last_change":"2026-03-10T11:32:09.497693+0000","last_active":"2026-03-10T11:32:14.095302+0000","last_peered":"2026-03-10T11:32:14.095302+0000","last_clean":"2026-03-10T11:32:14.095302+0000","last_became_active":"2026-03-10T11:32:09.497549+0000","last_became_peered":"2026-03-10T11:32:09.497549+0000","last_unstale":"2026-03-10T11:32:14.095302+0000","last_undegraded":"2026-03-10T11:32:14.095302+0000","last_fullsized":"2026-03-10T11:32:14.095302+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T23:07:22.489823+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,0,5],"acting":[2,0,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"6.8","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753123+0000","last_change":"2026-03-10T11:32:11.512386+0000","last_active":"2026-03-10T11:32:13.753123+0000","last_peered":"2026-03-10T11:32:13.753123+0000","last_clean":"2026-03-10T11:32:13.753123+0000","last_became_active":"2026-03-10T11:32:11.511538+0000","last_became_peered":"2026-03-10T11:32:11.511538+0000","last_unstale":"2026-03-10T11:32:13.753123+0000","last_undegraded":"2026-03-10T11:32:13.753123+0000","last_fullsized":"2026-03-10T11:32:13.753123+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:48:17.780752+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,2,3],"acting":[7,2,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"4.9","version":"55'12","reported_seq":46,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059256+0000","last_change":"2026-03-10T11:32:07.483133+0000","last_active":"2026-03-10T11:32:14.059256+0000","last_peered":"2026-03-10T11:32:14.059256+0000","last_clean":"2026-03-10T11:32:14.059256+0000","last_became_active":"2026-03-10T11:32:07.482407+0000","last_became_peered":"2026-03-10T11:32:07.482407+0000","last_unstale":"2026-03-10T11:32:14.059256+0000","last_undegraded":"2026-03-10T11:32:14.059256+0000","last_fullsized":"2026-03-10T11:32:14.059256+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":12,"log_dups_size":0,"ondisk_log_size":12,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:07:32.679946+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":220,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":25,"num_read_kb":16,"num_write":14,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,1,3],"acting":[4,1,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.e","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753020+0000","last_change":"2026-03-10T11:32:05.479546+0000","last_active":"2026-03-10T11:32:13.753020+0000","last_peered":"2026-03-10T11:32:13.753020+0000","last_clean":"2026-03-10T11:32:13.753020+0000","last_became_active":"2026-03-10T11:32:05.479050+0000","last_became_peered":"2026-03-10T11:32:05.479050+0000","last_unstale":"2026-03-10T11:32:13.753020+0000","last_undegraded":"2026-03-10T11:32:13.753020+0000","last_fullsized":"2026-03-10T11:32:13.753020+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:33:49.505723+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,1],"acting":[7,4,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.8","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.095213+0000","last_change":"2026-03-10T11:32:09.498834+0000","last_active":"2026-03-10T11:32:14.095213+0000","last_peered":"2026-03-10T11:32:14.095213+0000","last_clean":"2026-03-10T11:32:14.095213+0000","last_became_active":"2026-03-10T11:32:09.498594+0000","last_became_peered":"2026-03-10T11:32:09.498594+0000","last_unstale":"2026-03-10T11:32:14.095213+0000","last_undegraded":"2026-03-10T11:32:14.095213+0000","last_fullsized":"2026-03-10T11:32:14.095213+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:10:52.238205+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,0,1],"acting":[2,0,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"6.b","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.433102+0000","last_change":"2026-03-10T11:32:11.510427+0000","last_active":"2026-03-10T11:32:13.433102+0000","last_peered":"2026-03-10T11:32:13.433102+0000","last_clean":"2026-03-10T11:32:13.433102+0000","last_became_active":"2026-03-10T11:32:11.510285+0000","last_became_peered":"2026-03-10T11:32:11.510285+0000","last_unstale":"2026-03-10T11:32:13.433102+0000","last_undegraded":"2026-03-10T11:32:13.433102+0000","last_fullsized":"2026-03-10T11:32:13.433102+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:33:36.994107+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,1],"acting":[3,7,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.8","version":"55'15","reported_seq":48,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434844+0000","last_change":"2026-03-10T11:32:07.597775+0000","last_active":"2026-03-10T11:32:13.434844+0000","last_peered":"2026-03-10T11:32:13.434844+0000","last_clean":"2026-03-10T11:32:13.434844+0000","last_became_active":"2026-03-10T11:32:07.597555+0000","last_became_peered":"2026-03-10T11:32:07.597555+0000","last_unstale":"2026-03-10T11:32:13.434844+0000","last_undegraded":"2026-03-10T11:32:13.434844+0000","last_fullsized":"2026-03-10T11:32:13.434844+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":15,"log_dups_size":0,"ondisk_log_size":15,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:55:39.053348+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":7,"num_object_clones":0,"num_object_copies":21,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":7,"num_whiteouts":0,"num_read":26,"num_read_kb":17,"num_write":16,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,7,6],"acting":[5,7,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"3.f","version":"48'2","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752919+0000","last_change":"2026-03-10T11:32:05.484022+0000","last_active":"2026-03-10T11:32:13.752919+0000","last_peered":"2026-03-10T11:32:13.752919+0000","last_clean":"2026-03-10T11:32:13.752919+0000","last_became_active":"2026-03-10T11:32:05.483921+0000","last_became_peered":"2026-03-10T11:32:05.483921+0000","last_unstale":"2026-03-10T11:32:13.752919+0000","last_undegraded":"2026-03-10T11:32:13.752919+0000","last_fullsized":"2026-03-10T11:32:13.752919+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:52:04.652862+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":92,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":10,"num_read_kb":10,"num_write":4,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,0],"acting":[7,4,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.9","version":"55'8","reported_seq":30,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752990+0000","last_change":"2026-03-10T11:32:09.531130+0000","last_active":"2026-03-10T11:32:13.752990+0000","last_peered":"2026-03-10T11:32:13.752990+0000","last_clean":"2026-03-10T11:32:13.752990+0000","last_became_active":"2026-03-10T11:32:09.530942+0000","last_became_peered":"2026-03-10T11:32:09.530942+0000","last_unstale":"2026-03-10T11:32:13.752990+0000","last_undegraded":"2026-03-10T11:32:13.752990+0000","last_fullsized":"2026-03-10T11:32:13.752990+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:21:44.911594+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,6,4],"acting":[7,6,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"6.a","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.434856+0000","last_change":"2026-03-10T11:32:11.573996+0000","last_active":"2026-03-10T11:32:13.434856+0000","last_peered":"2026-03-10T11:32:13.434856+0000","last_clean":"2026-03-10T11:32:13.434856+0000","last_became_active":"2026-03-10T11:32:11.573890+0000","last_became_peered":"2026-03-10T11:32:11.573890+0000","last_unstale":"2026-03-10T11:32:13.434856+0000","last_undegraded":"2026-03-10T11:32:13.434856+0000","last_fullsized":"2026-03-10T11:32:13.434856+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:55:47.737007+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,6,0],"acting":[5,6,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"3.10","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431436+0000","last_change":"2026-03-10T11:32:05.480882+0000","last_active":"2026-03-10T11:32:13.431436+0000","last_peered":"2026-03-10T11:32:13.431436+0000","last_clean":"2026-03-10T11:32:13.431436+0000","last_became_active":"2026-03-10T11:32:05.480785+0000","last_became_peered":"2026-03-10T11:32:05.480785+0000","last_unstale":"2026-03-10T11:32:13.431436+0000","last_undegraded":"2026-03-10T11:32:13.431436+0000","last_fullsized":"2026-03-10T11:32:13.431436+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:58:13.892454+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,0,5],"acting":[6,0,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"4.17","version":"55'6","reported_seq":32,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432998+0000","last_change":"2026-03-10T11:32:07.598865+0000","last_active":"2026-03-10T11:32:13.432998+0000","last_peered":"2026-03-10T11:32:13.432998+0000","last_clean":"2026-03-10T11:32:13.432998+0000","last_became_active":"2026-03-10T11:32:07.598687+0000","last_became_peered":"2026-03-10T11:32:07.598687+0000","last_unstale":"2026-03-10T11:32:13.432998+0000","last_undegraded":"2026-03-10T11:32:13.432998+0000","last_fullsized":"2026-03-10T11:32:13.432998+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":6,"log_dups_size":0,"ondisk_log_size":6,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:45:00.774145+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":3,"num_object_clones":0,"num_object_copies":9,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":3,"num_whiteouts":0,"num_read":9,"num_read_kb":6,"num_write":6,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,1],"acting":[3,7,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"5.16","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435693+0000","last_change":"2026-03-10T11:32:09.507650+0000","last_active":"2026-03-10T11:32:13.435693+0000","last_peered":"2026-03-10T11:32:13.435693+0000","last_clean":"2026-03-10T11:32:13.435693+0000","last_became_active":"2026-03-10T11:32:09.501191+0000","last_became_peered":"2026-03-10T11:32:09.501191+0000","last_unstale":"2026-03-10T11:32:13.435693+0000","last_undegraded":"2026-03-10T11:32:13.435693+0000","last_fullsized":"2026-03-10T11:32:13.435693+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:30:26.718870+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,3,1],"acting":[5,3,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.15","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.753513+0000","last_change":"2026-03-10T11:32:11.577353+0000","last_active":"2026-03-10T11:32:13.753513+0000","last_peered":"2026-03-10T11:32:13.753513+0000","last_clean":"2026-03-10T11:32:13.753513+0000","last_became_active":"2026-03-10T11:32:11.575570+0000","last_became_peered":"2026-03-10T11:32:11.575570+0000","last_unstale":"2026-03-10T11:32:13.753513+0000","last_undegraded":"2026-03-10T11:32:13.753513+0000","last_fullsized":"2026-03-10T11:32:13.753513+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:52:18.344523+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,6,4],"acting":[7,6,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"4.16","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539668+0000","last_change":"2026-03-10T11:32:07.597957+0000","last_active":"2026-03-10T11:32:13.539668+0000","last_peered":"2026-03-10T11:32:13.539668+0000","last_clean":"2026-03-10T11:32:13.539668+0000","last_became_active":"2026-03-10T11:32:07.597610+0000","last_became_peered":"2026-03-10T11:32:07.597610+0000","last_unstale":"2026-03-10T11:32:13.539668+0000","last_undegraded":"2026-03-10T11:32:13.539668+0000","last_fullsized":"2026-03-10T11:32:13.539668+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:44:45.379060+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,3,7],"acting":[0,3,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"3.11","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752279+0000","last_change":"2026-03-10T11:32:05.479640+0000","last_active":"2026-03-10T11:32:13.752279+0000","last_peered":"2026-03-10T11:32:13.752279+0000","last_clean":"2026-03-10T11:32:13.752279+0000","last_became_active":"2026-03-10T11:32:05.479282+0000","last_became_peered":"2026-03-10T11:32:05.479282+0000","last_unstale":"2026-03-10T11:32:13.752279+0000","last_undegraded":"2026-03-10T11:32:13.752279+0000","last_fullsized":"2026-03-10T11:32:13.752279+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:25:22.717010+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,6],"acting":[7,4,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.17","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432008+0000","last_change":"2026-03-10T11:32:09.503165+0000","last_active":"2026-03-10T11:32:13.432008+0000","last_peered":"2026-03-10T11:32:13.432008+0000","last_clean":"2026-03-10T11:32:13.432008+0000","last_became_active":"2026-03-10T11:32:09.503019+0000","last_became_peered":"2026-03-10T11:32:09.503019+0000","last_unstale":"2026-03-10T11:32:13.432008+0000","last_undegraded":"2026-03-10T11:32:13.432008+0000","last_fullsized":"2026-03-10T11:32:13.432008+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:58:40.889689+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,7],"acting":[3,1,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"6.14","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.094844+0000","last_change":"2026-03-10T11:32:11.508463+0000","last_active":"2026-03-10T11:32:14.094844+0000","last_peered":"2026-03-10T11:32:14.094844+0000","last_clean":"2026-03-10T11:32:14.094844+0000","last_became_active":"2026-03-10T11:32:11.508373+0000","last_became_peered":"2026-03-10T11:32:11.508373+0000","last_unstale":"2026-03-10T11:32:14.094844+0000","last_undegraded":"2026-03-10T11:32:14.094844+0000","last_fullsized":"2026-03-10T11:32:14.094844+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:11:49.345325+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,4,7],"acting":[2,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"4.15","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435565+0000","last_change":"2026-03-10T11:32:07.597830+0000","last_active":"2026-03-10T11:32:13.435565+0000","last_peered":"2026-03-10T11:32:13.435565+0000","last_clean":"2026-03-10T11:32:13.435565+0000","last_became_active":"2026-03-10T11:32:07.597705+0000","last_became_peered":"2026-03-10T11:32:07.597705+0000","last_unstale":"2026-03-10T11:32:13.435565+0000","last_undegraded":"2026-03-10T11:32:13.435565+0000","last_fullsized":"2026-03-10T11:32:13.435565+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:51:57.119228+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,7,3],"acting":[5,7,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"3.12","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.538926+0000","last_change":"2026-03-10T11:32:05.494103+0000","last_active":"2026-03-10T11:32:13.538926+0000","last_peered":"2026-03-10T11:32:13.538926+0000","last_clean":"2026-03-10T11:32:13.538926+0000","last_became_active":"2026-03-10T11:32:05.493873+0000","last_became_peered":"2026-03-10T11:32:05.493873+0000","last_unstale":"2026-03-10T11:32:13.538926+0000","last_undegraded":"2026-03-10T11:32:13.538926+0000","last_fullsized":"2026-03-10T11:32:13.538926+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:30:14.014213+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,7,3],"acting":[0,7,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"5.14","version":"55'8","reported_seq":30,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432766+0000","last_change":"2026-03-10T11:32:09.503097+0000","last_active":"2026-03-10T11:32:13.432766+0000","last_peered":"2026-03-10T11:32:13.432766+0000","last_clean":"2026-03-10T11:32:13.432766+0000","last_became_active":"2026-03-10T11:32:09.502897+0000","last_became_peered":"2026-03-10T11:32:09.502897+0000","last_unstale":"2026-03-10T11:32:13.432766+0000","last_undegraded":"2026-03-10T11:32:13.432766+0000","last_fullsized":"2026-03-10T11:32:13.432766+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:19:50.521963+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,2],"acting":[3,7,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"6.17","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058501+0000","last_change":"2026-03-10T11:32:11.508381+0000","last_active":"2026-03-10T11:32:14.058501+0000","last_peered":"2026-03-10T11:32:14.058501+0000","last_clean":"2026-03-10T11:32:14.058501+0000","last_became_active":"2026-03-10T11:32:11.508250+0000","last_became_peered":"2026-03-10T11:32:11.508250+0000","last_unstale":"2026-03-10T11:32:14.058501+0000","last_undegraded":"2026-03-10T11:32:14.058501+0000","last_fullsized":"2026-03-10T11:32:14.058501+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:44:59.324849+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,2,5],"acting":[4,2,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"4.14","version":"55'10","reported_seq":38,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432952+0000","last_change":"2026-03-10T11:32:07.598047+0000","last_active":"2026-03-10T11:32:13.432952+0000","last_peered":"2026-03-10T11:32:13.432952+0000","last_clean":"2026-03-10T11:32:13.432952+0000","last_became_active":"2026-03-10T11:32:07.596958+0000","last_became_peered":"2026-03-10T11:32:07.596958+0000","last_unstale":"2026-03-10T11:32:13.432952+0000","last_undegraded":"2026-03-10T11:32:13.432952+0000","last_fullsized":"2026-03-10T11:32:13.432952+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":10,"log_dups_size":0,"ondisk_log_size":10,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:41:42.255637+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":15,"num_read_kb":10,"num_write":10,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,7],"acting":[3,1,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.13","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752816+0000","last_change":"2026-03-10T11:32:05.480017+0000","last_active":"2026-03-10T11:32:13.752816+0000","last_peered":"2026-03-10T11:32:13.752816+0000","last_clean":"2026-03-10T11:32:13.752816+0000","last_became_active":"2026-03-10T11:32:05.479411+0000","last_became_peered":"2026-03-10T11:32:05.479411+0000","last_unstale":"2026-03-10T11:32:13.752816+0000","last_undegraded":"2026-03-10T11:32:13.752816+0000","last_fullsized":"2026-03-10T11:32:13.752816+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:31:09.439155+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,2],"acting":[7,4,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.15","version":"55'8","reported_seq":33,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435802+0000","last_change":"2026-03-10T11:32:09.508210+0000","last_active":"2026-03-10T11:32:13.435802+0000","last_peered":"2026-03-10T11:32:13.435802+0000","last_clean":"2026-03-10T11:32:13.435802+0000","last_became_active":"2026-03-10T11:32:09.508106+0000","last_became_peered":"2026-03-10T11:32:09.508106+0000","last_unstale":"2026-03-10T11:32:13.435802+0000","last_undegraded":"2026-03-10T11:32:13.435802+0000","last_fullsized":"2026-03-10T11:32:13.435802+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":8,"log_dups_size":0,"ondisk_log_size":8,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:26:43.229572+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1,0],"acting":[5,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"6.16","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539060+0000","last_change":"2026-03-10T11:32:11.506003+0000","last_active":"2026-03-10T11:32:13.539060+0000","last_peered":"2026-03-10T11:32:13.539060+0000","last_clean":"2026-03-10T11:32:13.539060+0000","last_became_active":"2026-03-10T11:32:11.505923+0000","last_became_peered":"2026-03-10T11:32:11.505923+0000","last_unstale":"2026-03-10T11:32:13.539060+0000","last_undegraded":"2026-03-10T11:32:13.539060+0000","last_fullsized":"2026-03-10T11:32:13.539060+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T12:55:53.821261+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,7,3],"acting":[0,7,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.13","version":"55'11","reported_seq":42,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.059558+0000","last_change":"2026-03-10T11:32:07.480647+0000","last_active":"2026-03-10T11:32:14.059558+0000","last_peered":"2026-03-10T11:32:14.059558+0000","last_clean":"2026-03-10T11:32:14.059558+0000","last_became_active":"2026-03-10T11:32:07.480488+0000","last_became_peered":"2026-03-10T11:32:07.480488+0000","last_unstale":"2026-03-10T11:32:14.059558+0000","last_undegraded":"2026-03-10T11:32:14.059558+0000","last_fullsized":"2026-03-10T11:32:14.059558+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":11,"log_dups_size":0,"ondisk_log_size":11,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T19:58:00.369153+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":20,"num_read_kb":13,"num_write":12,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,6,1],"acting":[4,6,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"3.14","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:14.058329+0000","last_change":"2026-03-10T11:32:05.492719+0000","last_active":"2026-03-10T11:32:14.058329+0000","last_peered":"2026-03-10T11:32:14.058329+0000","last_clean":"2026-03-10T11:32:14.058329+0000","last_became_active":"2026-03-10T11:32:05.492463+0000","last_became_peered":"2026-03-10T11:32:05.492463+0000","last_unstale":"2026-03-10T11:32:14.058329+0000","last_undegraded":"2026-03-10T11:32:14.058329+0000","last_fullsized":"2026-03-10T11:32:14.058329+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:29:31.841348+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,7,6],"acting":[4,7,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[]},{"pgid":"5.12","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757427+0000","last_change":"2026-03-10T11:32:09.511793+0000","last_active":"2026-03-10T11:32:13.757427+0000","last_peered":"2026-03-10T11:32:13.757427+0000","last_clean":"2026-03-10T11:32:13.757427+0000","last_became_active":"2026-03-10T11:32:09.509850+0000","last_became_peered":"2026-03-10T11:32:09.509850+0000","last_unstale":"2026-03-10T11:32:13.757427+0000","last_undegraded":"2026-03-10T11:32:13.757427+0000","last_fullsized":"2026-03-10T11:32:13.757427+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:23:58.904271+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,3],"acting":[1,5,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"6.11","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431948+0000","last_change":"2026-03-10T11:32:11.514919+0000","last_active":"2026-03-10T11:32:13.431948+0000","last_peered":"2026-03-10T11:32:13.431948+0000","last_clean":"2026-03-10T11:32:13.431948+0000","last_became_active":"2026-03-10T11:32:11.514837+0000","last_became_peered":"2026-03-10T11:32:11.514837+0000","last_unstale":"2026-03-10T11:32:13.431948+0000","last_undegraded":"2026-03-10T11:32:13.431948+0000","last_fullsized":"2026-03-10T11:32:13.431948+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:27:53.843660+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0,5],"acting":[3,0,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.12","version":"55'9","reported_seq":39,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.757383+0000","last_change":"2026-03-10T11:32:07.482374+0000","last_active":"2026-03-10T11:32:13.757383+0000","last_peered":"2026-03-10T11:32:13.757383+0000","last_clean":"2026-03-10T11:32:13.757383+0000","last_became_active":"2026-03-10T11:32:07.482274+0000","last_became_peered":"2026-03-10T11:32:07.482274+0000","last_unstale":"2026-03-10T11:32:13.757383+0000","last_undegraded":"2026-03-10T11:32:13.757383+0000","last_fullsized":"2026-03-10T11:32:13.757383+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":9,"log_dups_size":0,"ondisk_log_size":9,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T21:41:26.092888+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":4,"num_object_clones":0,"num_object_copies":12,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":17,"num_read_kb":11,"num_write":10,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,6,2],"acting":[1,6,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.15","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752290+0000","last_change":"2026-03-10T11:32:05.480988+0000","last_active":"2026-03-10T11:32:13.752290+0000","last_peered":"2026-03-10T11:32:13.752290+0000","last_clean":"2026-03-10T11:32:13.752290+0000","last_became_active":"2026-03-10T11:32:05.480882+0000","last_became_peered":"2026-03-10T11:32:05.480882+0000","last_unstale":"2026-03-10T11:32:13.752290+0000","last_undegraded":"2026-03-10T11:32:13.752290+0000","last_fullsized":"2026-03-10T11:32:13.752290+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:12:29.005725+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,3,4],"acting":[7,3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"5.13","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432116+0000","last_change":"2026-03-10T11:32:09.498920+0000","last_active":"2026-03-10T11:32:13.432116+0000","last_peered":"2026-03-10T11:32:13.432116+0000","last_clean":"2026-03-10T11:32:13.432116+0000","last_became_active":"2026-03-10T11:32:09.498815+0000","last_became_peered":"2026-03-10T11:32:09.498815+0000","last_unstale":"2026-03-10T11:32:13.432116+0000","last_undegraded":"2026-03-10T11:32:13.432116+0000","last_fullsized":"2026-03-10T11:32:13.432116+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T20:08:28.641280+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0,1],"acting":[3,0,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"6.10","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539907+0000","last_change":"2026-03-10T11:32:11.510088+0000","last_active":"2026-03-10T11:32:13.539907+0000","last_peered":"2026-03-10T11:32:13.539907+0000","last_clean":"2026-03-10T11:32:13.539907+0000","last_became_active":"2026-03-10T11:32:11.510003+0000","last_became_peered":"2026-03-10T11:32:11.510003+0000","last_unstale":"2026-03-10T11:32:13.539907+0000","last_undegraded":"2026-03-10T11:32:13.539907+0000","last_fullsized":"2026-03-10T11:32:13.539907+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:45:24.974948+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,5,1],"acting":[0,5,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"4.11","version":"55'11","reported_seq":42,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432908+0000","last_change":"2026-03-10T11:32:07.598238+0000","last_active":"2026-03-10T11:32:13.432908+0000","last_peered":"2026-03-10T11:32:13.432908+0000","last_clean":"2026-03-10T11:32:13.432908+0000","last_became_active":"2026-03-10T11:32:07.598143+0000","last_became_peered":"2026-03-10T11:32:07.598143+0000","last_unstale":"2026-03-10T11:32:13.432908+0000","last_undegraded":"2026-03-10T11:32:13.432908+0000","last_fullsized":"2026-03-10T11:32:13.432908+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":11,"log_dups_size":0,"ondisk_log_size":11,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:11:32.909846+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":20,"num_read_kb":13,"num_write":12,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,7,6],"acting":[3,7,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.16","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.435324+0000","last_change":"2026-03-10T11:32:05.472568+0000","last_active":"2026-03-10T11:32:13.435324+0000","last_peered":"2026-03-10T11:32:13.435324+0000","last_clean":"2026-03-10T11:32:13.435324+0000","last_became_active":"2026-03-10T11:32:05.472437+0000","last_became_peered":"2026-03-10T11:32:05.472437+0000","last_unstale":"2026-03-10T11:32:13.435324+0000","last_undegraded":"2026-03-10T11:32:13.435324+0000","last_fullsized":"2026-03-10T11:32:13.435324+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:52:02.801467+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,7,1],"acting":[5,7,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"5.10","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752768+0000","last_change":"2026-03-10T11:32:09.519265+0000","last_active":"2026-03-10T11:32:13.752768+0000","last_peered":"2026-03-10T11:32:13.752768+0000","last_clean":"2026-03-10T11:32:13.752768+0000","last_became_active":"2026-03-10T11:32:09.519124+0000","last_became_peered":"2026-03-10T11:32:09.519124+0000","last_unstale":"2026-03-10T11:32:13.752768+0000","last_undegraded":"2026-03-10T11:32:13.752768+0000","last_fullsized":"2026-03-10T11:32:13.752768+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:26:19.264009+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,4,6],"acting":[7,4,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"6.13","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432894+0000","last_change":"2026-03-10T11:32:11.573852+0000","last_active":"2026-03-10T11:32:13.432894+0000","last_peered":"2026-03-10T11:32:13.432894+0000","last_clean":"2026-03-10T11:32:13.432894+0000","last_became_active":"2026-03-10T11:32:11.573716+0000","last_became_peered":"2026-03-10T11:32:11.573716+0000","last_unstale":"2026-03-10T11:32:13.432894+0000","last_undegraded":"2026-03-10T11:32:13.432894+0000","last_fullsized":"2026-03-10T11:32:13.432894+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T17:33:18.220367+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0,6],"acting":[3,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.10","version":"55'4","reported_seq":29,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432725+0000","last_change":"2026-03-10T11:32:07.493099+0000","last_active":"2026-03-10T11:32:13.432725+0000","last_peered":"2026-03-10T11:32:13.432725+0000","last_clean":"2026-03-10T11:32:13.432725+0000","last_became_active":"2026-03-10T11:32:07.493016+0000","last_became_peered":"2026-03-10T11:32:07.493016+0000","last_unstale":"2026-03-10T11:32:13.432725+0000","last_undegraded":"2026-03-10T11:32:13.432725+0000","last_fullsized":"2026-03-10T11:32:13.432725+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":4,"log_dups_size":0,"ondisk_log_size":4,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:32:04.890080+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":6,"num_read_kb":4,"num_write":4,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,6],"acting":[3,1,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"3.17","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.538956+0000","last_change":"2026-03-10T11:32:05.481179+0000","last_active":"2026-03-10T11:32:13.538956+0000","last_peered":"2026-03-10T11:32:13.538956+0000","last_clean":"2026-03-10T11:32:13.538956+0000","last_became_active":"2026-03-10T11:32:05.481078+0000","last_became_peered":"2026-03-10T11:32:05.481078+0000","last_unstale":"2026-03-10T11:32:13.538956+0000","last_undegraded":"2026-03-10T11:32:13.538956+0000","last_fullsized":"2026-03-10T11:32:13.538956+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T16:20:19.773427+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,5,3],"acting":[0,5,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]},{"pgid":"5.11","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431859+0000","last_change":"2026-03-10T11:32:09.517608+0000","last_active":"2026-03-10T11:32:13.431859+0000","last_peered":"2026-03-10T11:32:13.431859+0000","last_clean":"2026-03-10T11:32:13.431859+0000","last_became_active":"2026-03-10T11:32:09.517388+0000","last_became_peered":"2026-03-10T11:32:09.517388+0000","last_unstale":"2026-03-10T11:32:13.431859+0000","last_undegraded":"2026-03-10T11:32:13.431859+0000","last_fullsized":"2026-03-10T11:32:13.431859+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:30:58.947416+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,4,7],"acting":[6,4,7],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"6.12","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.752434+0000","last_change":"2026-03-10T11:32:11.509710+0000","last_active":"2026-03-10T11:32:13.752434+0000","last_peered":"2026-03-10T11:32:13.752434+0000","last_clean":"2026-03-10T11:32:13.752434+0000","last_became_active":"2026-03-10T11:32:11.508854+0000","last_became_peered":"2026-03-10T11:32:11.508854+0000","last_unstale":"2026-03-10T11:32:13.752434+0000","last_undegraded":"2026-03-10T11:32:13.752434+0000","last_fullsized":"2026-03-10T11:32:13.752434+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T15:36:56.911178+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,2,4],"acting":[7,2,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"6.1d","version":"0'0","reported_seq":15,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.756579+0000","last_change":"2026-03-10T11:32:11.509797+0000","last_active":"2026-03-10T11:32:13.756579+0000","last_peered":"2026-03-10T11:32:13.756579+0000","last_clean":"2026-03-10T11:32:13.756579+0000","last_became_active":"2026-03-10T11:32:11.509718+0000","last_became_peered":"2026-03-10T11:32:11.509718+0000","last_unstale":"2026-03-10T11:32:13.756579+0000","last_undegraded":"2026-03-10T11:32:13.756579+0000","last_fullsized":"2026-03-10T11:32:13.756579+0000","mapping_epoch":53,"log_start":"0'0","ondisk_log_start":"0'0","created":53,"last_epoch_clean":54,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:10.472887+0000","last_clean_scrub_stamp":"2026-03-10T11:32:10.472887+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T22:40:19.756646+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5,4],"acting":[1,5,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"3.18","version":"0'0","reported_seq":27,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.432542+0000","last_change":"2026-03-10T11:32:05.488107+0000","last_active":"2026-03-10T11:32:13.432542+0000","last_peered":"2026-03-10T11:32:13.432542+0000","last_clean":"2026-03-10T11:32:13.432542+0000","last_became_active":"2026-03-10T11:32:05.487957+0000","last_became_peered":"2026-03-10T11:32:05.487957+0000","last_unstale":"2026-03-10T11:32:13.432542+0000","last_undegraded":"2026-03-10T11:32:13.432542+0000","last_fullsized":"2026-03-10T11:32:13.432542+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:04.444639+0000","last_clean_scrub_stamp":"2026-03-10T11:32:04.444639+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T18:30:30.807507+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0,1],"acting":[3,0,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"4.1f","version":"55'11","reported_seq":42,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.431653+0000","last_change":"2026-03-10T11:32:07.492632+0000","last_active":"2026-03-10T11:32:13.431653+0000","last_peered":"2026-03-10T11:32:13.431653+0000","last_clean":"2026-03-10T11:32:13.431653+0000","last_became_active":"2026-03-10T11:32:07.492509+0000","last_became_peered":"2026-03-10T11:32:07.492509+0000","last_unstale":"2026-03-10T11:32:13.431653+0000","last_undegraded":"2026-03-10T11:32:13.431653+0000","last_fullsized":"2026-03-10T11:32:13.431653+0000","mapping_epoch":49,"log_start":"0'0","ondisk_log_start":"0'0","created":49,"last_epoch_clean":50,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:06.459549+0000","last_clean_scrub_stamp":"2026-03-10T11:32:06.459549+0000","objects_scrubbed":0,"log_size":11,"log_dups_size":0,"ondisk_log_size":11,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:55:48.589620+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":110,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":20,"num_read_kb":13,"num_write":12,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,5,1],"acting":[6,5,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]},{"pgid":"5.1e","version":"0'0","reported_seq":19,"reported_epoch":56,"state":"active+clean","last_fresh":"2026-03-10T11:32:13.539543+0000","last_change":"2026-03-10T11:32:09.505710+0000","last_active":"2026-03-10T11:32:13.539543+0000","last_peered":"2026-03-10T11:32:13.539543+0000","last_clean":"2026-03-10T11:32:13.539543+0000","last_became_active":"2026-03-10T11:32:09.503635+0000","last_became_peered":"2026-03-10T11:32:09.503635+0000","last_unstale":"2026-03-10T11:32:13.539543+0000","last_undegraded":"2026-03-10T11:32:13.539543+0000","last_fullsized":"2026-03-10T11:32:13.539543+0000","mapping_epoch":51,"log_start":"0'0","ondisk_log_start":"0'0","created":51,"last_epoch_clean":52,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T11:32:08.465007+0000","last_clean_scrub_stamp":"2026-03-10T11:32:08.465007+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:44:14.135459+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,7,2],"acting":[0,7,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[]}],"pool_stats":[{"poolid":6,"num_pg":32,"stat_sum":{"num_bytes":416,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":3,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1248,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":2,"ondisk_log_size":2,"up":96,"acting":96,"num_store_stats":8},{"poolid":5,"num_pg":32,"stat_sum":{"num_bytes":0,"num_objects":8,"num_object_clones":0,"num_object_copies":24,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":8,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":64,"ondisk_log_size":64,"up":96,"acting":96,"num_store_stats":8},{"poolid":4,"num_pg":32,"stat_sum":{"num_bytes":3702,"num_objects":178,"num_object_clones":0,"num_object_copies":534,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":178,"num_whiteouts":0,"num_read":698,"num_read_kb":455,"num_write":417,"num_write_kb":34,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":417792,"data_stored":11106,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":393,"ondisk_log_size":393,"up":96,"acting":96,"num_store_stats":8},{"poolid":3,"num_pg":32,"stat_sum":{"num_bytes":1613,"num_objects":6,"num_object_clones":0,"num_object_copies":18,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":6,"num_whiteouts":0,"num_read":24,"num_read_kb":24,"num_write":10,"num_write_kb":6,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":73728,"data_stored":4839,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":6,"ondisk_log_size":6,"up":96,"acting":96,"num_store_stats":8},{"poolid":2,"num_pg":3,"stat_sum":{"num_bytes":408,"num_objects":3,"num_object_clones":0,"num_object_copies":9,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":3,"num_whiteouts":0,"num_read":8,"num_read_kb":3,"num_write":6,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1224,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":8,"ondisk_log_size":8,"up":9,"acting":9,"num_store_stats":7},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":459280,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2777088,"data_stored":2755680,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":7}],"osd_stats":[{"osd":7,"up_from":44,"seq":188978561029,"num_pgs":53,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27904,"kb_used_data":1072,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939520,"statfs":{"total":21470642176,"available":21442068480,"internally_reserved":0,"allocated":1097728,"data_stored":719765,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":39,"seq":167503724551,"num_pgs":43,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27888,"kb_used_data":1052,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939536,"statfs":{"total":21470642176,"available":21442084864,"internally_reserved":0,"allocated":1077248,"data_stored":718123,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":32,"seq":137438953482,"num_pgs":47,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27444,"kb_used_data":604,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939980,"statfs":{"total":21470642176,"available":21442539520,"internally_reserved":0,"allocated":618496,"data_stored":258779,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":27,"seq":115964117004,"num_pgs":58,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27480,"kb_used_data":632,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939944,"statfs":{"total":21470642176,"available":21442502656,"internally_reserved":0,"allocated":647168,"data_stored":260441,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":22,"seq":94489280526,"num_pgs":56,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27472,"kb_used_data":636,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939952,"statfs":{"total":21470642176,"available":21442510848,"internally_reserved":0,"allocated":651264,"data_stored":259311,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":16,"seq":68719476752,"num_pgs":36,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27436,"kb_used_data":596,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939988,"statfs":{"total":21470642176,"available":21442547712,"internally_reserved":0,"allocated":610304,"data_stored":259548,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":12,"seq":51539607570,"num_pgs":57,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27504,"kb_used_data":668,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939920,"statfs":{"total":21470642176,"available":21442478080,"internally_reserved":0,"allocated":684032,"data_stored":260641,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738388,"num_pgs":46,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27912,"kb_used_data":1072,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939512,"statfs":{"total":21470642176,"available":21442060288,"internally_reserved":0,"allocated":1097728,"data_stored":718481,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":408,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":12288,"data_stored":138,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":12288,"data_stored":528,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":993,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":46,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1613,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":436,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":3,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":12288,"data_stored":1085,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":49152,"data_stored":1320,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":90112,"data_stored":2338,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":32768,"data_stored":798,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":73728,"data_stored":1898,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":53248,"data_stored":1474,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":36864,"data_stored":990,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":36864,"data_stored":1034,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":4,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":45056,"data_stored":1254,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":5,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":13,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":403,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":13,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":416,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":6,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":403,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T11:32:20.826 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-10T11:32:20.826 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-10T11:32:20.826 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-10T11:32:20.826 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph health --format=json 2026-03-10T11:32:21.042 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:32:21.306 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:32:21.307 INFO:teuthology.orchestra.run.vm06.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-10T11:32:21.319 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:21.319 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:21.319 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[49228]: from='client.24572 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T11:32:21.319 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:21.319 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-10T11:32:21.319 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[49228]: pgmap v112: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 4.5 KiB/s wr, 141 op/s 2026-03-10T11:32:21.319 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[49228]: from='client.14646 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T11:32:21.319 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:21.319 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:21.319 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[57405]: from='client.24572 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T11:32:21.319 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:21.320 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-10T11:32:21.320 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[57405]: pgmap v112: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 4.5 KiB/s wr, 141 op/s 2026-03-10T11:32:21.320 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:21 vm06 ceph-mon[57405]: from='client.14646 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T11:32:21.388 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-10T11:32:21.388 INFO:tasks.cephadm:Setup complete, yielding 2026-03-10T11:32:21.388 INFO:teuthology.run_tasks:Running task workunit... 2026-03-10T11:32:21.392 INFO:tasks.workunit:Pulling workunits from ref 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b 2026-03-10T11:32:21.392 INFO:tasks.workunit:Making a separate scratch dir for every client... 2026-03-10T11:32:21.392 DEBUG:teuthology.orchestra.run.vm06:> stat -- /home/ubuntu/cephtest/mnt.0 2026-03-10T11:32:21.414 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T11:32:21.415 INFO:teuthology.orchestra.run.vm06.stderr:stat: cannot statx '/home/ubuntu/cephtest/mnt.0': No such file or directory 2026-03-10T11:32:21.415 DEBUG:teuthology.orchestra.run.vm06:> mkdir -- /home/ubuntu/cephtest/mnt.0 2026-03-10T11:32:21.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:21 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:21.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:21 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:21.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:21 vm09 ceph-mon[54793]: from='client.24572 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T11:32:21.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:21 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' 2026-03-10T11:32:21.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:21 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-10T11:32:21.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:21 vm09 ceph-mon[54793]: pgmap v112: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 4.5 KiB/s wr, 141 op/s 2026-03-10T11:32:21.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:21 vm09 ceph-mon[54793]: from='client.14646 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T11:32:21.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:21 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: ignoring --setuser ceph since I am not root 2026-03-10T11:32:21.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:21 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: ignoring --setgroup ceph since I am not root 2026-03-10T11:32:21.481 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:21 vm09 ceph-mgr[56397]: -- 192.168.123.109:0/922789257 <== mon.1 v2:192.168.123.109:3300/0 4 ==== auth_reply(proto 2 0 (0) Success) ==== 194+0+0 (secure 0 0 0) 0x55694c7e14a0 con 0x55694c7be800 2026-03-10T11:32:21.481 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:21 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:21.384+0000 7fdebf919140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T11:32:21.481 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:21 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:21.426+0000 7fdebf919140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T11:32:21.485 INFO:tasks.workunit:Created dir /home/ubuntu/cephtest/mnt.0 2026-03-10T11:32:21.485 DEBUG:teuthology.orchestra.run.vm06:> cd -- /home/ubuntu/cephtest/mnt.0 && mkdir -- client.0 2026-03-10T11:32:21.548 INFO:tasks.workunit:timeout=1h 2026-03-10T11:32:21.548 INFO:tasks.workunit:cleanup=True 2026-03-10T11:32:21.548 DEBUG:teuthology.orchestra.run.vm06:> rm -rf /home/ubuntu/cephtest/clone.client.0 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.0 && cd /home/ubuntu/cephtest/clone.client.0 && git checkout 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b 2026-03-10T11:32:21.611 INFO:tasks.workunit.client.0.vm06.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.0'... 2026-03-10T11:32:21.781 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:21 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ignoring --setuser ceph since I am not root 2026-03-10T11:32:21.781 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:21 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ignoring --setgroup ceph since I am not root 2026-03-10T11:32:21.781 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:21 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:21.416+0000 7fe45f448140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T11:32:21.781 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:21 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:21.460+0000 7fe45f448140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T11:32:22.189 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:21 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:21.852+0000 7fdebf919140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T11:32:22.244 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:21 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:21.891+0000 7fe45f448140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T11:32:22.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:22 vm09 ceph-mon[54793]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-10T11:32:22.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:22 vm09 ceph-mon[54793]: mgrmap e17: y(active, since 2m), standbys: x 2026-03-10T11:32:22.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:22 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/782533611' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T11:32:22.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:22 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:22.188+0000 7fdebf919140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T11:32:22.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:22 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T11:32:22.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:22 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T11:32:22.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:22 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: from numpy import show_config as show_numpy_config 2026-03-10T11:32:22.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:22 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:22.282+0000 7fdebf919140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T11:32:22.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:22 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:22.322+0000 7fdebf919140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T11:32:22.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:22 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:22.396+0000 7fdebf919140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:22 vm06 ceph-mon[49228]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:22 vm06 ceph-mon[49228]: mgrmap e17: y(active, since 2m), standbys: x 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:22 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/782533611' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:22 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:22.244+0000 7fe45f448140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:22 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:22 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:22 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: from numpy import show_config as show_numpy_config 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:22 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:22.334+0000 7fe45f448140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:22 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:22.374+0000 7fe45f448140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:22 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:22.445+0000 7fe45f448140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:22 vm06 ceph-mon[57405]: from='mgr.14150 192.168.123.106:0/3006357255' entity='mgr.y' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:22 vm06 ceph-mon[57405]: mgrmap e17: y(active, since 2m), standbys: x 2026-03-10T11:32:22.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:22 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/782533611' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T11:32:23.185 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:22 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:22.915+0000 7fdebf919140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T11:32:23.185 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:23.024+0000 7fdebf919140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:32:23.185 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:23.065+0000 7fdebf919140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T11:32:23.185 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:23.104+0000 7fdebf919140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T11:32:23.186 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:23.146+0000 7fdebf919140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T11:32:23.242 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:22 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:22.968+0000 7fe45f448140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T11:32:23.242 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:23 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:23.082+0000 7fe45f448140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:32:23.242 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:23 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:23.123+0000 7fe45f448140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T11:32:23.242 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:23 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:23.159+0000 7fe45f448140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T11:32:23.242 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:23 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:23.202+0000 7fe45f448140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T11:32:23.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:23.184+0000 7fdebf919140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T11:32:23.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:23.366+0000 7fdebf919140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T11:32:23.480 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:23.416+0000 7fdebf919140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T11:32:23.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:23 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:23.242+0000 7fe45f448140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T11:32:23.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:23 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:23.420+0000 7fe45f448140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T11:32:23.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:23 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:23.473+0000 7fe45f448140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T11:32:23.945 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:23.652+0000 7fdebf919140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T11:32:24.015 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:23 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:23.713+0000 7fe45f448140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T11:32:24.230 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:23.943+0000 7fdebf919140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T11:32:24.230 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:23.984+0000 7fdebf919140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T11:32:24.230 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:24.027+0000 7fdebf919140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T11:32:24.230 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:24.107+0000 7fdebf919140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T11:32:24.230 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:24.147+0000 7fdebf919140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T11:32:24.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:24 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:24.015+0000 7fe45f448140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T11:32:24.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:24 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:24.060+0000 7fe45f448140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T11:32:24.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:24 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:24.102+0000 7fe45f448140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T11:32:24.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:24 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:24.185+0000 7fe45f448140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T11:32:24.281 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:24 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:24.226+0000 7fe45f448140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T11:32:24.504 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:24.233+0000 7fdebf919140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T11:32:24.504 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:24.355+0000 7fdebf919140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:32:24.600 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:24 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:24.323+0000 7fe45f448140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T11:32:24.600 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:24 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:24.448+0000 7fe45f448140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T11:32:24.971 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:24 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:24.600+0000 7fe45f448140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T11:32:24.971 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:24 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:24.641+0000 7fe45f448140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T11:32:24.971 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:24.502+0000 7fdebf919140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T11:32:24.971 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:24.544+0000 7fdebf919140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T11:32:24.971 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: [10/Mar/2026:11:32:24] ENGINE Bus STARTING 2026-03-10T11:32:24.971 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: CherryPy Checker: 2026-03-10T11:32:24.971 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: The Application mounted at '' has an empty config. 2026-03-10T11:32:24.971 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: 2026-03-10T11:32:24.971 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: [10/Mar/2026:11:32:24] ENGINE Serving on http://:::9283 2026-03-10T11:32:24.971 INFO:journalctl@ceph.mgr.x.vm09.stdout:Mar 10 11:32:24 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-x[56381]: [10/Mar/2026:11:32:24] ENGINE Bus STARTED 2026-03-10T11:32:24.973 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[49228]: Standby manager daemon x restarted 2026-03-10T11:32:24.973 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[49228]: Standby manager daemon x started 2026-03-10T11:32:24.973 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[49228]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/crt"}]: dispatch 2026-03-10T11:32:24.973 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[49228]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T11:32:24.973 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[49228]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/key"}]: dispatch 2026-03-10T11:32:24.973 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[49228]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T11:32:25.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:24 vm09 ceph-mon[54793]: Standby manager daemon x restarted 2026-03-10T11:32:25.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:24 vm09 ceph-mon[54793]: Standby manager daemon x started 2026-03-10T11:32:25.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:24 vm09 ceph-mon[54793]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/crt"}]: dispatch 2026-03-10T11:32:25.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:24 vm09 ceph-mon[54793]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T11:32:25.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:24 vm09 ceph-mon[54793]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/key"}]: dispatch 2026-03-10T11:32:25.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:24 vm09 ceph-mon[54793]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T11:32:25.236 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:25 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:25] ENGINE Bus STARTING 2026-03-10T11:32:25.236 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:25 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: CherryPy Checker: 2026-03-10T11:32:25.236 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:25 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: The Application mounted at '' has an empty config. 2026-03-10T11:32:25.236 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:25 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: 2026-03-10T11:32:25.236 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:25 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:25] ENGINE Serving on http://:::9283 2026-03-10T11:32:25.237 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[57405]: Standby manager daemon x restarted 2026-03-10T11:32:25.237 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[57405]: Standby manager daemon x started 2026-03-10T11:32:25.237 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[57405]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/crt"}]: dispatch 2026-03-10T11:32:25.237 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[57405]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T11:32:25.237 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[57405]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/x/key"}]: dispatch 2026-03-10T11:32:25.237 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:24 vm06 ceph-mon[57405]: from='mgr.? 192.168.123.109:0/2357926310' entity='mgr.x' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T11:32:25.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:25 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:25] ENGINE Bus STARTED 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: mgrmap e18: y(active, since 2m), standbys: x 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: Active manager daemon y restarted 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: Activating manager daemon y 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: osdmap e57: 8 total, 8 up, 8 in 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: mgrmap e19: y(active, starting, since 0.0153576s), standbys: x 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: Manager daemon y is now available 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:32:26.231 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:32:26.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:26.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:32:26.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:32:26.232 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:32:26.232 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:26 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: mgrmap e18: y(active, since 2m), standbys: x 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: Active manager daemon y restarted 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: Activating manager daemon y 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: osdmap e57: 8 total, 8 up, 8 in 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: mgrmap e19: y(active, starting, since 0.0153576s), standbys: x 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: Manager daemon y is now available 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:26.247 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:32:26.248 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:32:26.248 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: mgrmap e18: y(active, since 2m), standbys: x 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: Active manager daemon y restarted 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: Activating manager daemon y 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: osdmap e57: 8 total, 8 up, 8 in 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: mgrmap e19: y(active, starting, since 0.0153576s), standbys: x 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "y", "id": "y"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: Manager daemon y is now available 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/mirror_snapshot_schedule"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:32:26.249 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/y/trash_purge_schedule"}]: dispatch 2026-03-10T11:32:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[49228]: mgrmap e20: y(active, since 1.03977s), standbys: x 2026-03-10T11:32:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[49228]: pgmap v3: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[57405]: mgrmap e20: y(active, since 1.03977s), standbys: x 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[57405]: pgmap v3: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:27 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:27 vm09 ceph-mon[54793]: mgrmap e20: y(active, since 1.03977s), standbys: x 2026-03-10T11:32:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:27 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:27 vm09 ceph-mon[54793]: pgmap v3: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:27 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:27 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:27 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:27 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:27 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:27 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:27 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:27 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: [10/Mar/2026:11:32:26] ENGINE Bus STARTING 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: [10/Mar/2026:11:32:26] ENGINE Serving on https://192.168.123.106:7150 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: [10/Mar/2026:11:32:26] ENGINE Client ('192.168.123.106', 59790) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: [10/Mar/2026:11:32:26] ENGINE Serving on http://192.168.123.106:8765 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: [10/Mar/2026:11:32:26] ENGINE Bus STARTED 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: pgmap v4: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: mgrmap e21: y(active, since 2s), standbys: x 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:28 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: [10/Mar/2026:11:32:26] ENGINE Bus STARTING 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: [10/Mar/2026:11:32:26] ENGINE Serving on https://192.168.123.106:7150 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: [10/Mar/2026:11:32:26] ENGINE Client ('192.168.123.106', 59790) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: [10/Mar/2026:11:32:26] ENGINE Serving on http://192.168.123.106:8765 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: [10/Mar/2026:11:32:26] ENGINE Bus STARTED 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: pgmap v4: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: mgrmap e21: y(active, since 2s), standbys: x 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: [10/Mar/2026:11:32:26] ENGINE Bus STARTING 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: [10/Mar/2026:11:32:26] ENGINE Serving on https://192.168.123.106:7150 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: [10/Mar/2026:11:32:26] ENGINE Client ('192.168.123.106', 59790) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: [10/Mar/2026:11:32:26] ENGINE Serving on http://192.168.123.106:8765 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: [10/Mar/2026:11:32:26] ENGINE Bus STARTED 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: pgmap v4: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: mgrmap e21: y(active, since 2s), standbys: x 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:28.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:28 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:29.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:29 vm09 ceph-mon[54793]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:32:29.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:29 vm09 ceph-mon[54793]: Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:32:29.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:29 vm09 ceph-mon[54793]: Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:32:29.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:29 vm09 ceph-mon[54793]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:32:29.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:29 vm09 ceph-mon[54793]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:32:29.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:29 vm09 ceph-mon[54793]: Deploying daemon alertmanager.a on vm06 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[49228]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[49228]: Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[49228]: Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[49228]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[49228]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[49228]: Deploying daemon alertmanager.a on vm06 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[57405]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.conf 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[57405]: Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[57405]: Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[57405]: Updating vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[57405]: Updating vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/config/ceph.client.admin.keyring 2026-03-10T11:32:29.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:29 vm06 ceph-mon[57405]: Deploying daemon alertmanager.a on vm06 2026-03-10T11:32:30.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:30 vm09 ceph-mon[54793]: pgmap v5: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:30.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:30 vm09 ceph-mon[54793]: mgrmap e22: y(active, since 4s), standbys: x 2026-03-10T11:32:30.498 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:30 vm06 ceph-mon[49228]: pgmap v5: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:30.498 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:30 vm06 ceph-mon[49228]: mgrmap e22: y(active, since 4s), standbys: x 2026-03-10T11:32:30.498 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:30 vm06 ceph-mon[57405]: pgmap v5: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:30.498 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:30 vm06 ceph-mon[57405]: mgrmap e22: y(active, since 4s), standbys: x 2026-03-10T11:32:30.753 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:30 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:32:30] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:32:31.236 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 systemd[1]: Starting Ceph alertmanager.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:32:31.531 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:31] ENGINE Bus STOPPING 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 podman[86922]: 2026-03-10 11:32:31.137695398 +0000 UTC m=+0.011452759 image pull c8568f914cd25b2062c44e9f79f9c18da6e3b85fe0c47a12a2191c61426c2b19 quay.io/prometheus/alertmanager:v0.25.0 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 podman[86922]: 2026-03-10 11:32:31.238914337 +0000 UTC m=+0.112671687 volume create 276c777a39590941ccba6155b4cd8b907a457841e4a1fedb30db4959fb50e3b0 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 podman[86922]: 2026-03-10 11:32:31.244396889 +0000 UTC m=+0.118154239 container create a392601a12a81aa54a722e8b98624ab9788ef048e4e9125c07c51a7c41c8a6e4 (image=quay.io/prometheus/alertmanager:v0.25.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 podman[86922]: 2026-03-10 11:32:31.276904478 +0000 UTC m=+0.150661839 container init a392601a12a81aa54a722e8b98624ab9788ef048e4e9125c07c51a7c41c8a6e4 (image=quay.io/prometheus/alertmanager:v0.25.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 podman[86922]: 2026-03-10 11:32:31.279121458 +0000 UTC m=+0.152878808 container start a392601a12a81aa54a722e8b98624ab9788ef048e4e9125c07c51a7c41c8a6e4 (image=quay.io/prometheus/alertmanager:v0.25.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 bash[86922]: a392601a12a81aa54a722e8b98624ab9788ef048e4e9125c07c51a7c41c8a6e4 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 systemd[1]: Started Ceph alertmanager.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[86933]: ts=2026-03-10T11:32:31.299Z caller=main.go:240 level=info msg="Starting Alertmanager" version="(version=0.25.0, branch=HEAD, revision=258fab7cdd551f2cf251ed0348f0ad7289aee789)" 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[86933]: ts=2026-03-10T11:32:31.299Z caller=main.go:241 level=info build_context="(go=go1.19.4, user=root@abe866dd5717, date=20221222-14:51:36)" 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[86933]: ts=2026-03-10T11:32:31.300Z caller=cluster.go:185 level=info component=cluster msg="setting advertise address explicitly" addr=192.168.123.106 port=9094 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[86933]: ts=2026-03-10T11:32:31.302Z caller=cluster.go:681 level=info component=cluster msg="Waiting for gossip to settle..." interval=2s 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[86933]: ts=2026-03-10T11:32:31.333Z caller=coordinator.go:113 level=info component=configuration msg="Loading configuration file" file=/etc/alertmanager/alertmanager.yml 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[86933]: ts=2026-03-10T11:32:31.333Z caller=coordinator.go:126 level=info component=configuration msg="Completed loading of configuration file" file=/etc/alertmanager/alertmanager.yml 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[86933]: ts=2026-03-10T11:32:31.335Z caller=tls_config.go:232 level=info msg="Listening on" address=[::]:9093 2026-03-10T11:32:31.531 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[86933]: ts=2026-03-10T11:32:31.335Z caller=tls_config.go:235 level=info msg="TLS is disabled." http2=false address=[::]:9093 2026-03-10T11:32:32.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:31] ENGINE HTTP Server cherrypy._cpwsgi_server.CPWSGIServer(('::', 9283)) shut down 2026-03-10T11:32:32.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:31] ENGINE Bus STOPPED 2026-03-10T11:32:32.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:31] ENGINE Bus STARTING 2026-03-10T11:32:32.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:31] ENGINE Serving on http://:::9283 2026-03-10T11:32:32.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:31 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:31] ENGINE Bus STARTED 2026-03-10T11:32:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:32 vm09 ceph-mon[54793]: pgmap v6: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:32 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:32 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:32 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:32 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:32 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:32 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:32 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T11:32:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:32 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[57405]: pgmap v6: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T11:32:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[49228]: pgmap v6: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:32:32.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:32.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T11:32:32.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:32 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:33.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:33 vm09 ceph-mon[54793]: Regenerating cephadm self-signed grafana TLS certificates 2026-03-10T11:32:33.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:33 vm09 ceph-mon[54793]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T11:32:33.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:33 vm09 ceph-mon[54793]: Deploying daemon grafana.a on vm09 2026-03-10T11:32:33.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:33 vm06 ceph-mon[49228]: Regenerating cephadm self-signed grafana TLS certificates 2026-03-10T11:32:33.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:33 vm06 ceph-mon[49228]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T11:32:33.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:33 vm06 ceph-mon[49228]: Deploying daemon grafana.a on vm09 2026-03-10T11:32:33.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:33 vm06 ceph-mon[57405]: Regenerating cephadm self-signed grafana TLS certificates 2026-03-10T11:32:33.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:33 vm06 ceph-mon[57405]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T11:32:33.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:33 vm06 ceph-mon[57405]: Deploying daemon grafana.a on vm09 2026-03-10T11:32:33.781 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:33 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[86933]: ts=2026-03-10T11:32:33.303Z caller=cluster.go:706 level=info component=cluster msg="gossip not settled" polls=0 before=0 now=1 elapsed=2.000779804s 2026-03-10T11:32:34.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:34 vm09 ceph-mon[54793]: pgmap v7: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 26 KiB/s rd, 0 B/s wr, 11 op/s 2026-03-10T11:32:34.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:34 vm06 ceph-mon[49228]: pgmap v7: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 26 KiB/s rd, 0 B/s wr, 11 op/s 2026-03-10T11:32:34.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:34 vm06 ceph-mon[57405]: pgmap v7: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 26 KiB/s rd, 0 B/s wr, 11 op/s 2026-03-10T11:32:36.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:32:36.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:36 vm06 ceph-mon[49228]: pgmap v8: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 0 B/s wr, 8 op/s 2026-03-10T11:32:36.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:36 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:36.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:36 vm06 ceph-mon[57405]: pgmap v8: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 0 B/s wr, 8 op/s 2026-03-10T11:32:36.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:36 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:36 vm09 ceph-mon[54793]: pgmap v8: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 0 B/s wr, 8 op/s 2026-03-10T11:32:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:36 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:37.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:37 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:37.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:37 vm06 ceph-mon[49228]: pgmap v9: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 17 KiB/s rd, 0 B/s wr, 7 op/s 2026-03-10T11:32:37.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:37 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:37.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:37 vm06 ceph-mon[57405]: pgmap v9: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 17 KiB/s rd, 0 B/s wr, 7 op/s 2026-03-10T11:32:37.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:37 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:37.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:37 vm09 ceph-mon[54793]: pgmap v9: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 17 KiB/s rd, 0 B/s wr, 7 op/s 2026-03-10T11:32:39.404 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 systemd[1]: Starting Ceph grafana.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 podman[81007]: 2026-03-10 11:32:39.403725788 +0000 UTC m=+0.019584117 container create affeafdf638bc1051dd0ce4943c9d40f1f98b582f0657130cc9d85e63b42c3ac (image=quay.io/ceph/grafana:10.4.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a, maintainer=Grafana Labs ) 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 podman[81007]: 2026-03-10 11:32:39.448514241 +0000 UTC m=+0.064372570 container init affeafdf638bc1051dd0ce4943c9d40f1f98b582f0657130cc9d85e63b42c3ac (image=quay.io/ceph/grafana:10.4.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a, maintainer=Grafana Labs ) 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 podman[81007]: 2026-03-10 11:32:39.452097959 +0000 UTC m=+0.067956288 container start affeafdf638bc1051dd0ce4943c9d40f1f98b582f0657130cc9d85e63b42c3ac (image=quay.io/ceph/grafana:10.4.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a, maintainer=Grafana Labs ) 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 bash[81007]: affeafdf638bc1051dd0ce4943c9d40f1f98b582f0657130cc9d85e63b42c3ac 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 podman[81007]: 2026-03-10 11:32:39.395287353 +0000 UTC m=+0.011145682 image pull c8b91775d855b99270fc5d22f3c6737e8cca01ef4c25c8b0362295e0746fa39b quay.io/ceph/grafana:10.4.0 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 systemd[1]: Started Ceph grafana.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.557963904Z level=info msg="Starting Grafana" version=10.4.0 commit=03f502a94d17f7dc4e6c34acdf8428aedd986e4c branch=HEAD compiled=2026-03-10T11:32:39Z 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558534211Z level=info msg="Config loaded from" file=/usr/share/grafana/conf/defaults.ini 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558542637Z level=info msg="Config loaded from" file=/etc/grafana/grafana.ini 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558545703Z level=info msg="Config overridden from command line" arg="default.paths.data=/var/lib/grafana" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558549931Z level=info msg="Config overridden from command line" arg="default.paths.logs=/var/log/grafana" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558553177Z level=info msg="Config overridden from command line" arg="default.paths.plugins=/var/lib/grafana/plugins" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.55855531Z level=info msg="Config overridden from command line" arg="default.paths.provisioning=/etc/grafana/provisioning" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558557365Z level=info msg="Config overridden from command line" arg="default.log.mode=console" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558559498Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_DATA=/var/lib/grafana" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558561533Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_LOGS=/var/log/grafana" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558772667Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_PLUGINS=/var/lib/grafana/plugins" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558780011Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_PROVISIONING=/etc/grafana/provisioning" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558782706Z level=info msg=Target target=[all] 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558786714Z level=info msg="Path Home" path=/usr/share/grafana 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558788988Z level=info msg="Path Data" path=/var/lib/grafana 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558888956Z level=info msg="Path Logs" path=/var/log/grafana 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558893955Z level=info msg="Path Plugins" path=/var/lib/grafana/plugins 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558896109Z level=info msg="Path Provisioning" path=/etc/grafana/provisioning 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=settings t=2026-03-10T11:32:39.558900016Z level=info msg="App mode production" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=sqlstore t=2026-03-10T11:32:39.560492617Z level=info msg="Connecting to DB" dbtype=sqlite3 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=sqlstore t=2026-03-10T11:32:39.560506442Z level=warn msg="SQLite database file has broader permissions than it should" path=/var/lib/grafana/grafana.db mode=-rw-r--r-- expected=-rw-r----- 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.561201965Z level=info msg="Starting DB migrations" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.561961256Z level=info msg="Executing migration" id="create migration_log table" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.562530541Z level=info msg="Migration successfully executed" id="create migration_log table" duration=568.924µs 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.563191579Z level=info msg="Executing migration" id="create user table" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.563640799Z level=info msg="Migration successfully executed" id="create user table" duration=448.96µs 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.564182774Z level=info msg="Executing migration" id="add unique index user.login" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.564587482Z level=info msg="Migration successfully executed" id="add unique index user.login" duration=404.548µs 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.565173298Z level=info msg="Executing migration" id="add unique index user.email" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.565553399Z level=info msg="Migration successfully executed" id="add unique index user.email" duration=380.903µs 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.566147261Z level=info msg="Executing migration" id="drop index UQE_user_login - v1" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.566526761Z level=info msg="Migration successfully executed" id="drop index UQE_user_login - v1" duration=379.48µs 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.567112648Z level=info msg="Executing migration" id="drop index UQE_user_email - v1" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.56755144Z level=info msg="Migration successfully executed" id="drop index UQE_user_email - v1" duration=436.367µs 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.568414835Z level=info msg="Executing migration" id="Rename table user to user_v1 - v1" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.569689121Z level=info msg="Migration successfully executed" id="Rename table user to user_v1 - v1" duration=1.274426ms 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.570336983Z level=info msg="Executing migration" id="create user table v2" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.570804478Z level=info msg="Migration successfully executed" id="create user table v2" duration=468.999µs 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.571337346Z level=info msg="Executing migration" id="create index UQE_user_login - v2" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.571816091Z level=info msg="Migration successfully executed" id="create index UQE_user_login - v2" duration=482.964µs 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.572329262Z level=info msg="Executing migration" id="create index UQE_user_email - v2" 2026-03-10T11:32:39.656 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.572775398Z level=info msg="Migration successfully executed" id="create index UQE_user_email - v2" duration=446.306µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.573344303Z level=info msg="Executing migration" id="copy data_source v1 to v2" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.573694037Z level=info msg="Migration successfully executed" id="copy data_source v1 to v2" duration=347.44µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.574226322Z level=info msg="Executing migration" id="Drop old table user_v1" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.574622985Z level=info msg="Migration successfully executed" id="Drop old table user_v1" duration=396.904µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.575152046Z level=info msg="Executing migration" id="Add column help_flags1 to user table" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.575785492Z level=info msg="Migration successfully executed" id="Add column help_flags1 to user table" duration=634.057µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.576328548Z level=info msg="Executing migration" id="Update user table charset" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.576401555Z level=info msg="Migration successfully executed" id="Update user table charset" duration=73.417µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.577018449Z level=info msg="Executing migration" id="Add last_seen_at column to user" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.577627649Z level=info msg="Migration successfully executed" id="Add last_seen_at column to user" duration=609.541µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.578128267Z level=info msg="Executing migration" id="Add missing user data" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.578316799Z level=info msg="Migration successfully executed" id="Add missing user data" duration=188.032µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.578944805Z level=info msg="Executing migration" id="Add is_disabled column to user" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.579557381Z level=info msg="Migration successfully executed" id="Add is_disabled column to user" duration=612.637µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.580093826Z level=info msg="Executing migration" id="Add index user.login/user.email" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.58051299Z level=info msg="Migration successfully executed" id="Add index user.login/user.email" duration=419.275µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.581053983Z level=info msg="Executing migration" id="Add is_service_account column to user" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.581589104Z level=info msg="Migration successfully executed" id="Add is_service_account column to user" duration=534.039µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.582083229Z level=info msg="Executing migration" id="Update is_service_account column to nullable" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.585116787Z level=info msg="Migration successfully executed" id="Update is_service_account column to nullable" duration=3.032055ms 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.5856955Z level=info msg="Executing migration" id="Add uid column to user" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.586219181Z level=info msg="Migration successfully executed" id="Add uid column to user" duration=523.55µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.586674041Z level=info msg="Executing migration" id="Update uid column values for users" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.586813895Z level=info msg="Migration successfully executed" id="Update uid column values for users" duration=139.882µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.5873036Z level=info msg="Executing migration" id="Add unique index user_uid" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.587716613Z level=info msg="Migration successfully executed" id="Add unique index user_uid" duration=412.823µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.588250283Z level=info msg="Executing migration" id="create temp user table v1-7" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.588665109Z level=info msg="Migration successfully executed" id="create temp user table v1-7" duration=414.617µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.58918894Z level=info msg="Executing migration" id="create index IDX_temp_user_email - v1-7" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.589558591Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_email - v1-7" duration=369.692µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.590092831Z level=info msg="Executing migration" id="create index IDX_temp_user_org_id - v1-7" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.590457944Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_org_id - v1-7" duration=364.943µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.591036407Z level=info msg="Executing migration" id="create index IDX_temp_user_code - v1-7" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.591404275Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_code - v1-7" duration=366.085µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.591950368Z level=info msg="Executing migration" id="create index IDX_temp_user_status - v1-7" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.592313016Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_status - v1-7" duration=362.609µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.592812262Z level=info msg="Executing migration" id="Update temp_user table charset" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.592848479Z level=info msg="Migration successfully executed" id="Update temp_user table charset" duration=36.158µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.593395172Z level=info msg="Executing migration" id="drop index IDX_temp_user_email - v1" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.593796203Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_email - v1" duration=401.802µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.594259159Z level=info msg="Executing migration" id="drop index IDX_temp_user_org_id - v1" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.594653528Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_org_id - v1" duration=394.379µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.595129118Z level=info msg="Executing migration" id="drop index IDX_temp_user_code - v1" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.595502297Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_code - v1" duration=373.63µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.596002883Z level=info msg="Executing migration" id="drop index IDX_temp_user_status - v1" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.596357578Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_status - v1" duration=355.445µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.596830252Z level=info msg="Executing migration" id="Rename table temp_user to temp_user_tmp_qwerty - v1" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.600271724Z level=info msg="Migration successfully executed" id="Rename table temp_user to temp_user_tmp_qwerty - v1" duration=3.436493ms 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.601019663Z level=info msg="Executing migration" id="create temp_user v2" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.601437155Z level=info msg="Migration successfully executed" id="create temp_user v2" duration=417.642µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.601949344Z level=info msg="Executing migration" id="create index IDX_temp_user_email - v2" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.602357548Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_email - v2" duration=408.055µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.602804224Z level=info msg="Executing migration" id="create index IDX_temp_user_org_id - v2" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.603165691Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_org_id - v2" duration=359.884µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.603679872Z level=info msg="Executing migration" id="create index IDX_temp_user_code - v2" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.604034527Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_code - v2" duration=355.205µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.60448007Z level=info msg="Executing migration" id="create index IDX_temp_user_status - v2" 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.604873617Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_status - v2" duration=393.427µs 2026-03-10T11:32:39.657 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.605408669Z level=info msg="Executing migration" id="copy temp_user v1 to v2" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.605640442Z level=info msg="Migration successfully executed" id="copy temp_user v1 to v2" duration=231.443µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.606065158Z level=info msg="Executing migration" id="drop temp_user_tmp_qwerty" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.606345573Z level=info msg="Migration successfully executed" id="drop temp_user_tmp_qwerty" duration=278.08µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.606786527Z level=info msg="Executing migration" id="Set created for temp users that will otherwise prematurely expire" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.607003585Z level=info msg="Migration successfully executed" id="Set created for temp users that will otherwise prematurely expire" duration=217.057µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.607508739Z level=info msg="Executing migration" id="create star table" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.607846912Z level=info msg="Migration successfully executed" id="create star table" duration=338.283µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.608320258Z level=info msg="Executing migration" id="add unique index star.user_id_dashboard_id" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.608715157Z level=info msg="Migration successfully executed" id="add unique index star.user_id_dashboard_id" duration=394.578µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.609183634Z level=info msg="Executing migration" id="create org table v1" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.609521767Z level=info msg="Migration successfully executed" id="create org table v1" duration=338.103µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.60998258Z level=info msg="Executing migration" id="create index UQE_org_name - v1" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.610334969Z level=info msg="Migration successfully executed" id="create index UQE_org_name - v1" duration=352.47µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.610797836Z level=info msg="Executing migration" id="create org_user table v1" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.611112304Z level=info msg="Migration successfully executed" id="create org_user table v1" duration=315.541µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.61153205Z level=info msg="Executing migration" id="create index IDX_org_user_org_id - v1" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.611924235Z level=info msg="Migration successfully executed" id="create index IDX_org_user_org_id - v1" duration=390.692µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.612373946Z level=info msg="Executing migration" id="create index UQE_org_user_org_id_user_id - v1" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.612764858Z level=info msg="Migration successfully executed" id="create index UQE_org_user_org_id_user_id - v1" duration=390.662µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.613278719Z level=info msg="Executing migration" id="create index IDX_org_user_user_id - v1" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.61445994Z level=info msg="Migration successfully executed" id="create index IDX_org_user_user_id - v1" duration=1.179548ms 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.631774198Z level=info msg="Executing migration" id="Update org table charset" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.631916063Z level=info msg="Migration successfully executed" id="Update org table charset" duration=144.861µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.632729697Z level=info msg="Executing migration" id="Update org_user table charset" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.632791492Z level=info msg="Migration successfully executed" id="Update org_user table charset" duration=62.096µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.633448252Z level=info msg="Executing migration" id="Migrate all Read Only Viewers to Viewers" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.633653876Z level=info msg="Migration successfully executed" id="Migrate all Read Only Viewers to Viewers" duration=205.404µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.634241596Z level=info msg="Executing migration" id="create dashboard table" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.634820319Z level=info msg="Migration successfully executed" id="create dashboard table" duration=578.432µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.635378264Z level=info msg="Executing migration" id="add index dashboard.account_id" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.635847081Z level=info msg="Migration successfully executed" id="add index dashboard.account_id" duration=468.987µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.636373908Z level=info msg="Executing migration" id="add unique index dashboard_account_id_slug" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.636800165Z level=info msg="Migration successfully executed" id="add unique index dashboard_account_id_slug" duration=426.057µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.637349965Z level=info msg="Executing migration" id="create dashboard_tag table" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.637695612Z level=info msg="Migration successfully executed" id="create dashboard_tag table" duration=345.767µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.638184627Z level=info msg="Executing migration" id="add unique index dashboard_tag.dasboard_id_term" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.638562134Z level=info msg="Migration successfully executed" id="add unique index dashboard_tag.dasboard_id_term" duration=377.507µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.639130177Z level=info msg="Executing migration" id="drop index UQE_dashboard_tag_dashboard_id_term - v1" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.639574137Z level=info msg="Migration successfully executed" id="drop index UQE_dashboard_tag_dashboard_id_term - v1" duration=441.736µs 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.640125721Z level=info msg="Executing migration" id="Rename table dashboard to dashboard_v1 - v1" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.642105606Z level=info msg="Migration successfully executed" id="Rename table dashboard to dashboard_v1 - v1" duration=1.979255ms 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.6426369Z level=info msg="Executing migration" id="create dashboard v2" 2026-03-10T11:32:39.658 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.643004478Z level=info msg="Migration successfully executed" id="create dashboard v2" duration=367.438µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.643476312Z level=info msg="Executing migration" id="create index IDX_dashboard_org_id - v2" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.643869658Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_org_id - v2" duration=393.007µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.644357851Z level=info msg="Executing migration" id="create index UQE_dashboard_org_id_slug - v2" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.644775193Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_org_id_slug - v2" duration=417.062µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.645247657Z level=info msg="Executing migration" id="copy dashboard v1 to v2" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.645450096Z level=info msg="Migration successfully executed" id="copy dashboard v1 to v2" duration=202.389µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.645916299Z level=info msg="Executing migration" id="drop table dashboard_v1" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.646380828Z level=info msg="Migration successfully executed" id="drop table dashboard_v1" duration=464.458µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.646935015Z level=info msg="Executing migration" id="alter dashboard.data to mediumtext v1" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.646996621Z level=info msg="Migration successfully executed" id="alter dashboard.data to mediumtext v1" duration=61.635µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.647525361Z level=info msg="Executing migration" id="Add column updated_by in dashboard - v2" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.648200013Z level=info msg="Migration successfully executed" id="Add column updated_by in dashboard - v2" duration=674.872µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.648725066Z level=info msg="Executing migration" id="Add column created_by in dashboard - v2" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.649349555Z level=info msg="Migration successfully executed" id="Add column created_by in dashboard - v2" duration=624.52µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.649832149Z level=info msg="Executing migration" id="Add column gnetId in dashboard" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.650456838Z level=info msg="Migration successfully executed" id="Add column gnetId in dashboard" duration=624.73µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.650909955Z level=info msg="Executing migration" id="Add index for gnetId in dashboard" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.65131832Z level=info msg="Migration successfully executed" id="Add index for gnetId in dashboard" duration=407.593µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.651824187Z level=info msg="Executing migration" id="Add column plugin_id in dashboard" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.652458534Z level=info msg="Migration successfully executed" id="Add column plugin_id in dashboard" duration=635.83µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.652930768Z level=info msg="Executing migration" id="Add index for plugin_id in dashboard" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.653308305Z level=info msg="Migration successfully executed" id="Add index for plugin_id in dashboard" duration=377.418µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.653791349Z level=info msg="Executing migration" id="Add index for dashboard_id in dashboard_tag" 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.654152836Z level=info msg="Migration successfully executed" id="Add index for dashboard_id in dashboard_tag" duration=361.367µs 2026-03-10T11:32:39.659 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.654631973Z level=info msg="Executing migration" id="Update dashboard table charset" 2026-03-10T11:32:39.904 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.654665596Z level=info msg="Migration successfully executed" id="Update dashboard table charset" duration=33.903µs 2026-03-10T11:32:39.904 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.658037256Z level=info msg="Executing migration" id="Update dashboard_tag table charset" 2026-03-10T11:32:39.904 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.658049289Z level=info msg="Migration successfully executed" id="Update dashboard_tag table charset" duration=13.074µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.658540418Z level=info msg="Executing migration" id="Add column folder_id in dashboard" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.659297976Z level=info msg="Migration successfully executed" id="Add column folder_id in dashboard" duration=757.108µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.659840893Z level=info msg="Executing migration" id="Add column isFolder in dashboard" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.660575347Z level=info msg="Migration successfully executed" id="Add column isFolder in dashboard" duration=734.434µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.661104397Z level=info msg="Executing migration" id="Add column has_acl in dashboard" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.661857578Z level=info msg="Migration successfully executed" id="Add column has_acl in dashboard" duration=753.451µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.66235644Z level=info msg="Executing migration" id="Add column uid in dashboard" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.66307157Z level=info msg="Migration successfully executed" id="Add column uid in dashboard" duration=714.969µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.663544866Z level=info msg="Executing migration" id="Update uid column values in dashboard" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.663709925Z level=info msg="Migration successfully executed" id="Update uid column values in dashboard" duration=165.139µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.664282065Z level=info msg="Executing migration" id="Add unique index dashboard_org_id_uid" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.664715597Z level=info msg="Migration successfully executed" id="Add unique index dashboard_org_id_uid" duration=432.219µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.665233445Z level=info msg="Executing migration" id="Remove unique index org_id_slug" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.665661598Z level=info msg="Migration successfully executed" id="Remove unique index org_id_slug" duration=426.43µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.666168025Z level=info msg="Executing migration" id="Update dashboard title length" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.666183123Z level=info msg="Migration successfully executed" id="Update dashboard title length" duration=14.536µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.666817221Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_title_folder_id" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.667250882Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_title_folder_id" duration=433.442µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.667776035Z level=info msg="Executing migration" id="create dashboard_provisioning" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.668123185Z level=info msg="Migration successfully executed" id="create dashboard_provisioning" duration=347.441µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.66857954Z level=info msg="Executing migration" id="Rename table dashboard_provisioning to dashboard_provisioning_tmp_qwerty - v1" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.670311241Z level=info msg="Migration successfully executed" id="Rename table dashboard_provisioning to dashboard_provisioning_tmp_qwerty - v1" duration=1.731321ms 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.670830312Z level=info msg="Executing migration" id="create dashboard_provisioning v2" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.671150431Z level=info msg="Migration successfully executed" id="create dashboard_provisioning v2" duration=320.048µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.671645868Z level=info msg="Executing migration" id="create index IDX_dashboard_provisioning_dashboard_id - v2" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.672013146Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_provisioning_dashboard_id - v2" duration=367.748µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.672457588Z level=info msg="Executing migration" id="create index IDX_dashboard_provisioning_dashboard_id_name - v2" 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.672826919Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_provisioning_dashboard_id_name - v2" duration=367.548µs 2026-03-10T11:32:39.905 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.673294525Z level=info msg="Executing migration" id="copy dashboard_provisioning v1 to v2" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.673467919Z level=info msg="Migration successfully executed" id="copy dashboard_provisioning v1 to v2" duration=173.184µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.673912931Z level=info msg="Executing migration" id="drop dashboard_provisioning_tmp_qwerty" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.674185793Z level=info msg="Migration successfully executed" id="drop dashboard_provisioning_tmp_qwerty" duration=274.705µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.674676741Z level=info msg="Executing migration" id="Add check_sum column" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.675370981Z level=info msg="Migration successfully executed" id="Add check_sum column" duration=694.55µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.675902836Z level=info msg="Executing migration" id="Add index for dashboard_title" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.676254325Z level=info msg="Migration successfully executed" id="Add index for dashboard_title" duration=351.298µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.676690931Z level=info msg="Executing migration" id="delete tags for deleted dashboards" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.676784096Z level=info msg="Migration successfully executed" id="delete tags for deleted dashboards" duration=93.485µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.677263243Z level=info msg="Executing migration" id="delete stars for deleted dashboards" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.677350125Z level=info msg="Migration successfully executed" id="delete stars for deleted dashboards" duration=87.023µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.677703246Z level=info msg="Executing migration" id="Add index for dashboard_is_folder" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.678050496Z level=info msg="Migration successfully executed" id="Add index for dashboard_is_folder" duration=346.93µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.678534602Z level=info msg="Executing migration" id="Add isPublic for dashboard" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.679249Z level=info msg="Migration successfully executed" id="Add isPublic for dashboard" duration=714.328µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.679781556Z level=info msg="Executing migration" id="create data_source table" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.680191374Z level=info msg="Migration successfully executed" id="create data_source table" duration=408.065µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.680684617Z level=info msg="Executing migration" id="add index data_source.account_id" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.681062024Z level=info msg="Migration successfully executed" id="add index data_source.account_id" duration=377.116µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.681497147Z level=info msg="Executing migration" id="add unique index data_source.account_id_name" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.681865097Z level=info msg="Migration successfully executed" id="add unique index data_source.account_id_name" duration=367.579µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.682300942Z level=info msg="Executing migration" id="drop index IDX_data_source_account_id - v1" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.682677627Z level=info msg="Migration successfully executed" id="drop index IDX_data_source_account_id - v1" duration=379.149µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.683136376Z level=info msg="Executing migration" id="drop index UQE_data_source_account_id_name - v1" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.683498944Z level=info msg="Migration successfully executed" id="drop index UQE_data_source_account_id_name - v1" duration=362.538µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.683957964Z level=info msg="Executing migration" id="Rename table data_source to data_source_v1 - v1" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.685737404Z level=info msg="Migration successfully executed" id="Rename table data_source to data_source_v1 - v1" duration=1.779261ms 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.686245635Z level=info msg="Executing migration" id="create data_source table v2" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.686688555Z level=info msg="Migration successfully executed" id="create data_source table v2" duration=442.6µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.687141332Z level=info msg="Executing migration" id="create index IDX_data_source_org_id - v2" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.68753591Z level=info msg="Migration successfully executed" id="create index IDX_data_source_org_id - v2" duration=395.911µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.688069389Z level=info msg="Executing migration" id="create index UQE_data_source_org_id_name - v2" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.688494014Z level=info msg="Migration successfully executed" id="create index UQE_data_source_org_id_name - v2" duration=424.705µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.689019658Z level=info msg="Executing migration" id="Drop old table data_source_v1 #2" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.689333545Z level=info msg="Migration successfully executed" id="Drop old table data_source_v1 #2" duration=313.947µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.68980608Z level=info msg="Executing migration" id="Add column with_credentials" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.690676068Z level=info msg="Migration successfully executed" id="Add column with_credentials" duration=869.879µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.69120076Z level=info msg="Executing migration" id="Add secure json data column" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.692042516Z level=info msg="Migration successfully executed" id="Add secure json data column" duration=842.427µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.692540137Z level=info msg="Executing migration" id="Update data_source table charset" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.692552551Z level=info msg="Migration successfully executed" id="Update data_source table charset" duration=12.594µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.693122898Z level=info msg="Executing migration" id="Update initial version to 1" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.693238595Z level=info msg="Migration successfully executed" id="Update initial version to 1" duration=115.887µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.693666205Z level=info msg="Executing migration" id="Add read_only data column" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.694542585Z level=info msg="Migration successfully executed" id="Add read_only data column" duration=876.3µs 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.695076274Z level=info msg="Executing migration" id="Migrate logging ds to loki ds" 2026-03-10T11:32:39.906 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.695179978Z level=info msg="Migration successfully executed" id="Migrate logging ds to loki ds" duration=103.584µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.695735689Z level=info msg="Executing migration" id="Update json_data with nulls" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.695845815Z level=info msg="Migration successfully executed" id="Update json_data with nulls" duration=111.99µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.696352484Z level=info msg="Executing migration" id="Add uid column" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.697204367Z level=info msg="Migration successfully executed" id="Add uid column" duration=851.844µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.697672995Z level=info msg="Executing migration" id="Update uid value" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.697773633Z level=info msg="Migration successfully executed" id="Update uid value" duration=100.829µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.698279911Z level=info msg="Executing migration" id="Add unique index datasource_org_id_uid" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.698673297Z level=info msg="Migration successfully executed" id="Add unique index datasource_org_id_uid" duration=393.215µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.699135974Z level=info msg="Executing migration" id="add unique index datasource_org_id_is_default" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.699526474Z level=info msg="Migration successfully executed" id="add unique index datasource_org_id_is_default" duration=390.631µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.700023764Z level=info msg="Executing migration" id="create api_key table" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.70040014Z level=info msg="Migration successfully executed" id="create api_key table" duration=376.294µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.700912147Z level=info msg="Executing migration" id="add index api_key.account_id" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.701270138Z level=info msg="Migration successfully executed" id="add index api_key.account_id" duration=356.197µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.701763291Z level=info msg="Executing migration" id="add index api_key.key" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.702111463Z level=info msg="Migration successfully executed" id="add index api_key.key" duration=347.751µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.70262349Z level=info msg="Executing migration" id="add index api_key.account_id_name" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.702993654Z level=info msg="Migration successfully executed" id="add index api_key.account_id_name" duration=370.083µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.703516252Z level=info msg="Executing migration" id="drop index IDX_api_key_account_id - v1" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.703916811Z level=info msg="Migration successfully executed" id="drop index IDX_api_key_account_id - v1" duration=400.409µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.704380559Z level=info msg="Executing migration" id="drop index UQE_api_key_key - v1" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.70475471Z level=info msg="Migration successfully executed" id="drop index UQE_api_key_key - v1" duration=373.83µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.705220572Z level=info msg="Executing migration" id="drop index UQE_api_key_account_id_name - v1" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.705585946Z level=info msg="Migration successfully executed" id="drop index UQE_api_key_account_id_name - v1" duration=365.584µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.706119104Z level=info msg="Executing migration" id="Rename table api_key to api_key_v1 - v1" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.708097457Z level=info msg="Migration successfully executed" id="Rename table api_key to api_key_v1 - v1" duration=1.979355ms 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.708720272Z level=info msg="Executing migration" id="create api_key table v2" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.709067192Z level=info msg="Migration successfully executed" id="create api_key table v2" duration=349.314µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.709554924Z level=info msg="Executing migration" id="create index IDX_api_key_org_id - v2" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.709930518Z level=info msg="Migration successfully executed" id="create index IDX_api_key_org_id - v2" duration=375.513µs 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.710412409Z level=info msg="Executing migration" id="create index UQE_api_key_key - v2" 2026-03-10T11:32:39.907 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.710772865Z level=info msg="Migration successfully executed" id="create index UQE_api_key_key - v2" duration=360.104µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.711221664Z level=info msg="Executing migration" id="create index UQE_api_key_org_id_name - v2" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.711576619Z level=info msg="Migration successfully executed" id="create index UQE_api_key_org_id_name - v2" duration=354.733µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.712021381Z level=info msg="Executing migration" id="copy api_key v1 to v2" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.712202039Z level=info msg="Migration successfully executed" id="copy api_key v1 to v2" duration=180.398µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.712691816Z level=info msg="Executing migration" id="Drop old table api_key_v1" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.712961641Z level=info msg="Migration successfully executed" id="Drop old table api_key_v1" duration=269.704µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.713400743Z level=info msg="Executing migration" id="Update api_key table charset" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.713412745Z level=info msg="Migration successfully executed" id="Update api_key table charset" duration=12.153µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.713912992Z level=info msg="Executing migration" id="Add expires to api_key table" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.714725523Z level=info msg="Migration successfully executed" id="Add expires to api_key table" duration=812.621µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.715157401Z level=info msg="Executing migration" id="Add service account foreign key" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.715970724Z level=info msg="Migration successfully executed" id="Add service account foreign key" duration=812.982µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.716425515Z level=info msg="Executing migration" id="set service account foreign key to nil if 0" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.716510143Z level=info msg="Migration successfully executed" id="set service account foreign key to nil if 0" duration=86.423µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.717028283Z level=info msg="Executing migration" id="Add last_used_at to api_key table" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.717948706Z level=info msg="Migration successfully executed" id="Add last_used_at to api_key table" duration=919.581µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.718432261Z level=info msg="Executing migration" id="Add is_revoked column to api_key table" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.71934028Z level=info msg="Migration successfully executed" id="Add is_revoked column to api_key table" duration=907.818µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.719882746Z level=info msg="Executing migration" id="create dashboard_snapshot table v4" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.720288154Z level=info msg="Migration successfully executed" id="create dashboard_snapshot table v4" duration=405.289µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.720768905Z level=info msg="Executing migration" id="drop table dashboard_snapshot_v4 #1" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.721075939Z level=info msg="Migration successfully executed" id="drop table dashboard_snapshot_v4 #1" duration=307.094µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.721536141Z level=info msg="Executing migration" id="create dashboard_snapshot table v5 #2" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.722065311Z level=info msg="Migration successfully executed" id="create dashboard_snapshot table v5 #2" duration=528.308µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.722590684Z level=info msg="Executing migration" id="create index UQE_dashboard_snapshot_key - v5" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.723062117Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_snapshot_key - v5" duration=471.132µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.723537677Z level=info msg="Executing migration" id="create index UQE_dashboard_snapshot_delete_key - v5" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.724000222Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_snapshot_delete_key - v5" duration=461.203µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.724462779Z level=info msg="Executing migration" id="create index IDX_dashboard_snapshot_user_id - v5" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.724914754Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_snapshot_user_id - v5" duration=451.785µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.725451239Z level=info msg="Executing migration" id="alter dashboard_snapshot to mediumtext v2" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.725479912Z level=info msg="Migration successfully executed" id="alter dashboard_snapshot to mediumtext v2" duration=30.758µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.726025984Z level=info msg="Executing migration" id="Update dashboard_snapshot table charset" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.726053305Z level=info msg="Migration successfully executed" id="Update dashboard_snapshot table charset" duration=27.862µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.726528635Z level=info msg="Executing migration" id="Add column external_delete_url to dashboard_snapshots table" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.727466731Z level=info msg="Migration successfully executed" id="Add column external_delete_url to dashboard_snapshots table" duration=937.325µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.728007775Z level=info msg="Executing migration" id="Add encrypted dashboard json column" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.728886298Z level=info msg="Migration successfully executed" id="Add encrypted dashboard json column" duration=877.932µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.729353954Z level=info msg="Executing migration" id="Change dashboard_encrypted column to MEDIUMBLOB" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.729380624Z level=info msg="Migration successfully executed" id="Change dashboard_encrypted column to MEDIUMBLOB" duration=27.251µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.729851655Z level=info msg="Executing migration" id="create quota table v1" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.730180691Z level=info msg="Migration successfully executed" id="create quota table v1" duration=329.017µs 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.730668704Z level=info msg="Executing migration" id="create index UQE_quota_org_id_user_id_target - v1" 2026-03-10T11:32:39.908 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.731024922Z level=info msg="Migration successfully executed" id="create index UQE_quota_org_id_user_id_target - v1" duration=356.508µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.731482487Z level=info msg="Executing migration" id="Update quota table charset" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.731493909Z level=info msg="Migration successfully executed" id="Update quota table charset" duration=11.592µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.732011549Z level=info msg="Executing migration" id="create plugin_setting table" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.732341406Z level=info msg="Migration successfully executed" id="create plugin_setting table" duration=329.677µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.732809271Z level=info msg="Executing migration" id="create index UQE_plugin_setting_org_id_plugin_id - v1" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.733190344Z level=info msg="Migration successfully executed" id="create index UQE_plugin_setting_org_id_plugin_id - v1" duration=380.793µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.733671264Z level=info msg="Executing migration" id="Add column plugin_version to plugin_settings" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.734548036Z level=info msg="Migration successfully executed" id="Add column plugin_version to plugin_settings" duration=876.691µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.735004791Z level=info msg="Executing migration" id="Update plugin_setting table charset" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.735016503Z level=info msg="Migration successfully executed" id="Update plugin_setting table charset" duration=12.133µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.735518722Z level=info msg="Executing migration" id="create session table" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.735919123Z level=info msg="Migration successfully executed" id="create session table" duration=400.23µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.736446258Z level=info msg="Executing migration" id="Drop old table playlist table" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.736485131Z level=info msg="Migration successfully executed" id="Drop old table playlist table" duration=39.063µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.736993603Z level=info msg="Executing migration" id="Drop old table playlist_item table" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.737055028Z level=info msg="Migration successfully executed" id="Drop old table playlist_item table" duration=60.513µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.737567056Z level=info msg="Executing migration" id="create playlist table v2" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.737954432Z level=info msg="Migration successfully executed" id="create playlist table v2" duration=387.165µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.738460519Z level=info msg="Executing migration" id="create playlist item table v2" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.738843426Z level=info msg="Migration successfully executed" id="create playlist item table v2" duration=382.998µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.739325277Z level=info msg="Executing migration" id="Update playlist table charset" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.739337129Z level=info msg="Migration successfully executed" id="Update playlist table charset" duration=13.786µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.739925451Z level=info msg="Executing migration" id="Update playlist_item table charset" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.739937473Z level=info msg="Migration successfully executed" id="Update playlist_item table charset" duration=13.535µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.740505216Z level=info msg="Executing migration" id="Add playlist column created_at" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.741873657Z level=info msg="Migration successfully executed" id="Add playlist column created_at" duration=1.36762ms 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.742356961Z level=info msg="Executing migration" id="Add playlist column updated_at" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.74333386Z level=info msg="Migration successfully executed" id="Add playlist column updated_at" duration=975.296µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.743816344Z level=info msg="Executing migration" id="drop preferences table v2" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.74385708Z level=info msg="Migration successfully executed" id="drop preferences table v2" duration=40.676µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.744356946Z level=info msg="Executing migration" id="drop preferences table v3" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.744392312Z level=info msg="Migration successfully executed" id="drop preferences table v3" duration=35.587µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.744807789Z level=info msg="Executing migration" id="create preferences table v3" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.745147956Z level=info msg="Migration successfully executed" id="create preferences table v3" duration=340.187µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.745618457Z level=info msg="Executing migration" id="Update preferences table charset" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.745632283Z level=info msg="Migration successfully executed" id="Update preferences table charset" duration=14.116µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.746117511Z level=info msg="Executing migration" id="Add column team_id in preferences" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.747085192Z level=info msg="Migration successfully executed" id="Add column team_id in preferences" duration=967.531µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.747541537Z level=info msg="Executing migration" id="Update team_id column values in preferences" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.747641464Z level=info msg="Migration successfully executed" id="Update team_id column values in preferences" duration=99.957µs 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.748099991Z level=info msg="Executing migration" id="Add column week_start in preferences" 2026-03-10T11:32:39.909 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.749084934Z level=info msg="Migration successfully executed" id="Add column week_start in preferences" duration=984.834µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.749538094Z level=info msg="Executing migration" id="Add column preferences.json_data" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.75050363Z level=info msg="Migration successfully executed" id="Add column preferences.json_data" duration=965.496µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.750999348Z level=info msg="Executing migration" id="alter preferences.json_data to mediumtext v1" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.751028473Z level=info msg="Migration successfully executed" id="alter preferences.json_data to mediumtext v1" duration=29.456µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.751574585Z level=info msg="Executing migration" id="Add preferences index org_id" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.752042151Z level=info msg="Migration successfully executed" id="Add preferences index org_id" duration=467.416µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.752518252Z level=info msg="Executing migration" id="Add preferences index user_id" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.752944078Z level=info msg="Migration successfully executed" id="Add preferences index user_id" duration=425.646µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.753429787Z level=info msg="Executing migration" id="create alert table v1" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.753919484Z level=info msg="Migration successfully executed" id="create alert table v1" duration=489.506µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.754479643Z level=info msg="Executing migration" id="add index alert org_id & id " 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.754929194Z level=info msg="Migration successfully executed" id="add index alert org_id & id " duration=449.121µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.755412819Z level=info msg="Executing migration" id="add index alert state" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.75580339Z level=info msg="Migration successfully executed" id="add index alert state" duration=390.18µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.756276106Z level=info msg="Executing migration" id="add index alert dashboard_id" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.756697483Z level=info msg="Migration successfully executed" id="add index alert dashboard_id" duration=421.277µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.757164759Z level=info msg="Executing migration" id="Create alert_rule_tag table v1" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.757466162Z level=info msg="Migration successfully executed" id="Create alert_rule_tag table v1" duration=301.434µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.757964206Z level=info msg="Executing migration" id="Add unique index alert_rule_tag.alert_id_tag_id" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.758368582Z level=info msg="Migration successfully executed" id="Add unique index alert_rule_tag.alert_id_tag_id" duration=405.959µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.758836628Z level=info msg="Executing migration" id="drop index UQE_alert_rule_tag_alert_id_tag_id - v1" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.759232729Z level=info msg="Migration successfully executed" id="drop index UQE_alert_rule_tag_alert_id_tag_id - v1" duration=397.023µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.75971963Z level=info msg="Executing migration" id="Rename table alert_rule_tag to alert_rule_tag_v1 - v1" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.762658722Z level=info msg="Migration successfully executed" id="Rename table alert_rule_tag to alert_rule_tag_v1 - v1" duration=2.93849ms 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.763180418Z level=info msg="Executing migration" id="Create alert_rule_tag table v2" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.763550561Z level=info msg="Migration successfully executed" id="Create alert_rule_tag table v2" duration=369.852µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.764124205Z level=info msg="Executing migration" id="create index UQE_alert_rule_tag_alert_id_tag_id - Add unique index alert_rule_tag.alert_id_tag_id V2" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.764546195Z level=info msg="Migration successfully executed" id="create index UQE_alert_rule_tag_alert_id_tag_id - Add unique index alert_rule_tag.alert_id_tag_id V2" duration=421.79µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.76505655Z level=info msg="Executing migration" id="copy alert_rule_tag v1 to v2" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.765234894Z level=info msg="Migration successfully executed" id="copy alert_rule_tag v1 to v2" duration=178.594µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.765650733Z level=info msg="Executing migration" id="drop table alert_rule_tag_v1" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.765941586Z level=info msg="Migration successfully executed" id="drop table alert_rule_tag_v1" duration=290.794µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.766452223Z level=info msg="Executing migration" id="create alert_notification table v1" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.766826563Z level=info msg="Migration successfully executed" id="create alert_notification table v1" duration=374.37µs 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.76728994Z level=info msg="Executing migration" id="Add column is_default" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.768428652Z level=info msg="Migration successfully executed" id="Add column is_default" duration=1.151516ms 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.768952682Z level=info msg="Executing migration" id="Add column frequency" 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.77008947Z level=info msg="Migration successfully executed" id="Add column frequency" duration=1.136668ms 2026-03-10T11:32:39.910 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.770554211Z level=info msg="Executing migration" id="Add column send_reminder" 2026-03-10T11:32:39.911 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.771806865Z level=info msg="Migration successfully executed" id="Add column send_reminder" duration=1.252344ms 2026-03-10T11:32:39.911 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.772277577Z level=info msg="Executing migration" id="Add column disable_resolve_message" 2026-03-10T11:32:39.911 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.773370231Z level=info msg="Migration successfully executed" id="Add column disable_resolve_message" duration=1.092325ms 2026-03-10T11:32:39.911 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.773878653Z level=info msg="Executing migration" id="add index alert_notification org_id & name" 2026-03-10T11:32:39.911 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.774263884Z level=info msg="Migration successfully executed" id="add index alert_notification org_id & name" duration=385.041µs 2026-03-10T11:32:39.911 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.774706943Z level=info msg="Executing migration" id="Update alert table charset" 2026-03-10T11:32:39.911 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.774719407Z level=info msg="Migration successfully executed" id="Update alert table charset" duration=12.853µs 2026-03-10T11:32:39.911 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.775232297Z level=info msg="Executing migration" id="Update alert_notification table charset" 2026-03-10T11:32:39.911 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.775243949Z level=info msg="Migration successfully executed" id="Update alert_notification table charset" duration=12.514µs 2026-03-10T11:32:39.911 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.775630702Z level=info msg="Executing migration" id="create notification_journal table v1" 2026-03-10T11:32:39.911 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.775952785Z level=info msg="Migration successfully executed" id="create notification_journal table v1" duration=322.113µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.776438865Z level=info msg="Executing migration" id="add index notification_journal org_id & alert_id & notifier_id" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.776830237Z level=info msg="Migration successfully executed" id="add index notification_journal org_id & alert_id & notifier_id" duration=391.673µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.777265703Z level=info msg="Executing migration" id="drop alert_notification_journal" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.777645974Z level=info msg="Migration successfully executed" id="drop alert_notification_journal" duration=380.472µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.778127025Z level=info msg="Executing migration" id="create alert_notification_state table v1" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.778471499Z level=info msg="Migration successfully executed" id="create alert_notification_state table v1" duration=345.716µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.778930037Z level=info msg="Executing migration" id="add index alert_notification_state org_id & alert_id & notifier_id" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.779309078Z level=info msg="Migration successfully executed" id="add index alert_notification_state org_id & alert_id & notifier_id" duration=378.89µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.779736036Z level=info msg="Executing migration" id="Add for to alert table" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.780861533Z level=info msg="Migration successfully executed" id="Add for to alert table" duration=1.124986ms 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.7812925Z level=info msg="Executing migration" id="Add column uid in alert_notification" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.782421232Z level=info msg="Migration successfully executed" id="Add column uid in alert_notification" duration=1.128422ms 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.782933822Z level=info msg="Executing migration" id="Update uid column values in alert_notification" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.783027057Z level=info msg="Migration successfully executed" id="Update uid column values in alert_notification" duration=93.074µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.783480455Z level=info msg="Executing migration" id="Add unique index alert_notification_org_id_uid" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.783886766Z level=info msg="Migration successfully executed" id="Add unique index alert_notification_org_id_uid" duration=405.049µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.784311381Z level=info msg="Executing migration" id="Remove unique index org_id_name" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.784712081Z level=info msg="Migration successfully executed" id="Remove unique index org_id_name" duration=400.421µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.785202037Z level=info msg="Executing migration" id="Add column secure_settings in alert_notification" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.786313869Z level=info msg="Migration successfully executed" id="Add column secure_settings in alert_notification" duration=1.112181ms 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.786777627Z level=info msg="Executing migration" id="alter alert.settings to mediumtext" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.786804808Z level=info msg="Migration successfully executed" id="alter alert.settings to mediumtext" duration=27.692µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.787306466Z level=info msg="Executing migration" id="Add non-unique index alert_notification_state_alert_id" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.78770331Z level=info msg="Migration successfully executed" id="Add non-unique index alert_notification_state_alert_id" duration=396.614µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.788130158Z level=info msg="Executing migration" id="Add non-unique index alert_rule_tag_alert_id" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.788662475Z level=info msg="Migration successfully executed" id="Add non-unique index alert_rule_tag_alert_id" duration=532.177µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.789177538Z level=info msg="Executing migration" id="Drop old annotation table v4" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.789218626Z level=info msg="Migration successfully executed" id="Drop old annotation table v4" duration=41.348µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.789768585Z level=info msg="Executing migration" id="create annotation table v5" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.790191306Z level=info msg="Migration successfully executed" id="create annotation table v5" duration=422.731µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.790682095Z level=info msg="Executing migration" id="add index annotation 0 v3" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.791100748Z level=info msg="Migration successfully executed" id="add index annotation 0 v3" duration=418.794µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.791635539Z level=info msg="Executing migration" id="add index annotation 1 v3" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.792051778Z level=info msg="Migration successfully executed" id="add index annotation 1 v3" duration=416.37µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.79250114Z level=info msg="Executing migration" id="add index annotation 2 v3" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.792926075Z level=info msg="Migration successfully executed" id="add index annotation 2 v3" duration=424.735µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.793383391Z level=info msg="Executing migration" id="add index annotation 3 v3" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.793809499Z level=info msg="Migration successfully executed" id="add index annotation 3 v3" duration=425.967µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.794297493Z level=info msg="Executing migration" id="add index annotation 4 v3" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.794749027Z level=info msg="Migration successfully executed" id="add index annotation 4 v3" duration=451.114µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.79521556Z level=info msg="Executing migration" id="Update annotation table charset" 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.795227714Z level=info msg="Migration successfully executed" id="Update annotation table charset" duration=12.124µs 2026-03-10T11:32:39.912 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.795736515Z level=info msg="Executing migration" id="Add column region_id to annotation table" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.797017363Z level=info msg="Migration successfully executed" id="Add column region_id to annotation table" duration=1.279314ms 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.797476242Z level=info msg="Executing migration" id="Drop category_id index" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.797876781Z level=info msg="Migration successfully executed" id="Drop category_id index" duration=398.866µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.798339759Z level=info msg="Executing migration" id="Add column tags to annotation table" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.799517242Z level=info msg="Migration successfully executed" id="Add column tags to annotation table" duration=1.177294ms 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.80007162Z level=info msg="Executing migration" id="Create annotation_tag table v2" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.800382642Z level=info msg="Migration successfully executed" id="Create annotation_tag table v2" duration=311.021µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.800863903Z level=info msg="Executing migration" id="Add unique index annotation_tag.annotation_id_tag_id" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.801265003Z level=info msg="Migration successfully executed" id="Add unique index annotation_tag.annotation_id_tag_id" duration=401.191µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.801692844Z level=info msg="Executing migration" id="drop index UQE_annotation_tag_annotation_id_tag_id - v2" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.802083406Z level=info msg="Migration successfully executed" id="drop index UQE_annotation_tag_annotation_id_tag_id - v2" duration=390.311µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.802546472Z level=info msg="Executing migration" id="Rename table annotation_tag to annotation_tag_v2 - v2" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.805710755Z level=info msg="Migration successfully executed" id="Rename table annotation_tag to annotation_tag_v2 - v2" duration=3.163872ms 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.806192707Z level=info msg="Executing migration" id="Create annotation_tag table v3" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.806504129Z level=info msg="Migration successfully executed" id="Create annotation_tag table v3" duration=311.182µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.806977947Z level=info msg="Executing migration" id="create index UQE_annotation_tag_annotation_id_tag_id - Add unique index annotation_tag.annotation_id_tag_id V3" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.80736983Z level=info msg="Migration successfully executed" id="create index UQE_annotation_tag_annotation_id_tag_id - Add unique index annotation_tag.annotation_id_tag_id V3" duration=390.56µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.807807309Z level=info msg="Executing migration" id="copy annotation_tag v2 to v3" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.807965175Z level=info msg="Migration successfully executed" id="copy annotation_tag v2 to v3" duration=157.655µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.808412131Z level=info msg="Executing migration" id="drop table annotation_tag_v2" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.808692806Z level=info msg="Migration successfully executed" id="drop table annotation_tag_v2" duration=280.345µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.809155592Z level=info msg="Executing migration" id="Update alert annotations and set TEXT to empty" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.809246032Z level=info msg="Migration successfully executed" id="Update alert annotations and set TEXT to empty" duration=90.459µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.809702627Z level=info msg="Executing migration" id="Add created time to annotation table" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.810898475Z level=info msg="Migration successfully executed" id="Add created time to annotation table" duration=1.196039ms 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.811310115Z level=info msg="Executing migration" id="Add updated time to annotation table" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.812499121Z level=info msg="Migration successfully executed" id="Add updated time to annotation table" duration=1.188756ms 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.813020637Z level=info msg="Executing migration" id="Add index for created in annotation table" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.813411819Z level=info msg="Migration successfully executed" id="Add index for created in annotation table" duration=390.932µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.813861551Z level=info msg="Executing migration" id="Add index for updated in annotation table" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.814266329Z level=info msg="Migration successfully executed" id="Add index for updated in annotation table" duration=404.516µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.81473683Z level=info msg="Executing migration" id="Convert existing annotations from seconds to milliseconds" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.814860581Z level=info msg="Migration successfully executed" id="Convert existing annotations from seconds to milliseconds" duration=123.591µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.815399439Z level=info msg="Executing migration" id="Add epoch_end column" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.816842571Z level=info msg="Migration successfully executed" id="Add epoch_end column" duration=1.444014ms 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.817324953Z level=info msg="Executing migration" id="Add index for epoch_end" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.817764687Z level=info msg="Migration successfully executed" id="Add index for epoch_end" duration=439.052µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.818206132Z level=info msg="Executing migration" id="Make epoch_end the same as epoch" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.81829049Z level=info msg="Migration successfully executed" id="Make epoch_end the same as epoch" duration=84.418µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.818756583Z level=info msg="Executing migration" id="Move region to single row" 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.81892048Z level=info msg="Migration successfully executed" id="Move region to single row" duration=163.946µs 2026-03-10T11:32:39.913 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.819393426Z level=info msg="Executing migration" id="Remove index org_id_epoch from annotation table" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.819806729Z level=info msg="Migration successfully executed" id="Remove index org_id_epoch from annotation table" duration=414.324µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.820243246Z level=info msg="Executing migration" id="Remove index org_id_dashboard_id_panel_id_epoch from annotation table" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.82064606Z level=info msg="Migration successfully executed" id="Remove index org_id_dashboard_id_panel_id_epoch from annotation table" duration=402.743µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.821093276Z level=info msg="Executing migration" id="Add index for org_id_dashboard_id_epoch_end_epoch on annotation table" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.821493566Z level=info msg="Migration successfully executed" id="Add index for org_id_dashboard_id_epoch_end_epoch on annotation table" duration=400.33µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.821969467Z level=info msg="Executing migration" id="Add index for org_id_epoch_end_epoch on annotation table" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.822351111Z level=info msg="Migration successfully executed" id="Add index for org_id_epoch_end_epoch on annotation table" duration=381.515µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.822811132Z level=info msg="Executing migration" id="Remove index org_id_epoch_epoch_end from annotation table" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.823193659Z level=info msg="Migration successfully executed" id="Remove index org_id_epoch_epoch_end from annotation table" duration=382.225µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.82365413Z level=info msg="Executing migration" id="Add index for alert_id on annotation table" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.824026818Z level=info msg="Migration successfully executed" id="Add index for alert_id on annotation table" duration=372.637µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.824455681Z level=info msg="Executing migration" id="Increase tags column to length 4096" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.824482561Z level=info msg="Migration successfully executed" id="Increase tags column to length 4096" duration=27.251µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.825017802Z level=info msg="Executing migration" id="create test_data table" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.825372647Z level=info msg="Migration successfully executed" id="create test_data table" duration=354.464µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.82587113Z level=info msg="Executing migration" id="create dashboard_version table v1" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.826208912Z level=info msg="Migration successfully executed" id="create dashboard_version table v1" duration=337.722µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.82668862Z level=info msg="Executing migration" id="add index dashboard_version.dashboard_id" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.827063292Z level=info msg="Migration successfully executed" id="add index dashboard_version.dashboard_id" duration=374.551µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.827526769Z level=info msg="Executing migration" id="add unique index dashboard_version.dashboard_id and dashboard_version.version" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.827930074Z level=info msg="Migration successfully executed" id="add unique index dashboard_version.dashboard_id and dashboard_version.version" duration=403.505µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.828405574Z level=info msg="Executing migration" id="Set dashboard version to 1 where 0" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.828496284Z level=info msg="Migration successfully executed" id="Set dashboard version to 1 where 0" duration=90.259µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.828994356Z level=info msg="Executing migration" id="save existing dashboard data in dashboard_version table v1" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.829171027Z level=info msg="Migration successfully executed" id="save existing dashboard data in dashboard_version table v1" duration=174.737µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.829565725Z level=info msg="Executing migration" id="alter dashboard_version.data to mediumtext v1" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.829607693Z level=info msg="Migration successfully executed" id="alter dashboard_version.data to mediumtext v1" duration=41.707µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.830081551Z level=info msg="Executing migration" id="create team table" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.830395668Z level=info msg="Migration successfully executed" id="create team table" duration=313.526µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.830839089Z level=info msg="Executing migration" id="add index team.org_id" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.83127767Z level=info msg="Migration successfully executed" id="add index team.org_id" duration=438.571µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.831789017Z level=info msg="Executing migration" id="add unique index team_org_id_name" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.832297899Z level=info msg="Migration successfully executed" id="add unique index team_org_id_name" duration=508.541µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.832793718Z level=info msg="Executing migration" id="Add column uid in team" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.834100403Z level=info msg="Migration successfully executed" id="Add column uid in team" duration=1.306084ms 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.834554483Z level=info msg="Executing migration" id="Update uid column values in team" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.834677895Z level=info msg="Migration successfully executed" id="Update uid column values in team" duration=123.061µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.835174303Z level=info msg="Executing migration" id="Add unique index team_org_id_uid" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.835588178Z level=info msg="Migration successfully executed" id="Add unique index team_org_id_uid" duration=413.883µs 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.836055122Z level=info msg="Executing migration" id="create team member table" 2026-03-10T11:32:39.914 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.836391952Z level=info msg="Migration successfully executed" id="create team member table" duration=337.791µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.836876399Z level=info msg="Executing migration" id="add index team_member.org_id" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.837268372Z level=info msg="Migration successfully executed" id="add index team_member.org_id" duration=392.654µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.837718486Z level=info msg="Executing migration" id="add unique index team_member_org_id_team_id_user_id" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.838106171Z level=info msg="Migration successfully executed" id="add unique index team_member_org_id_team_id_user_id" duration=387.545µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.83863406Z level=info msg="Executing migration" id="add index team_member.team_id" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.839069675Z level=info msg="Migration successfully executed" id="add index team_member.team_id" duration=434.402µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.83955343Z level=info msg="Executing migration" id="Add column email to team table" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.84120442Z level=info msg="Migration successfully executed" id="Add column email to team table" duration=1.650809ms 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.841783384Z level=info msg="Executing migration" id="Add column external to team_member table" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.843276278Z level=info msg="Migration successfully executed" id="Add column external to team_member table" duration=1.493476ms 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.843803796Z level=info msg="Executing migration" id="Add column permission to team_member table" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.845155475Z level=info msg="Migration successfully executed" id="Add column permission to team_member table" duration=1.349545ms 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.845672303Z level=info msg="Executing migration" id="create dashboard acl table" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.846149095Z level=info msg="Migration successfully executed" id="create dashboard acl table" duration=476.561µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.846708562Z level=info msg="Executing migration" id="add index dashboard_acl_dashboard_id" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.847249265Z level=info msg="Migration successfully executed" id="add index dashboard_acl_dashboard_id" duration=540.141µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.847772754Z level=info msg="Executing migration" id="add unique index dashboard_acl_dashboard_id_user_id" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.84828886Z level=info msg="Migration successfully executed" id="add unique index dashboard_acl_dashboard_id_user_id" duration=487.032µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.848875458Z level=info msg="Executing migration" id="add unique index dashboard_acl_dashboard_id_team_id" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.849344857Z level=info msg="Migration successfully executed" id="add unique index dashboard_acl_dashboard_id_team_id" duration=469.158µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.849844702Z level=info msg="Executing migration" id="add index dashboard_acl_user_id" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.850292381Z level=info msg="Migration successfully executed" id="add index dashboard_acl_user_id" duration=448.64µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.850845024Z level=info msg="Executing migration" id="add index dashboard_acl_team_id" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.851253349Z level=info msg="Migration successfully executed" id="add index dashboard_acl_team_id" duration=408.044µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.851726615Z level=info msg="Executing migration" id="add index dashboard_acl_org_id_role" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.852130752Z level=info msg="Migration successfully executed" id="add index dashboard_acl_org_id_role" duration=404.859µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.852624586Z level=info msg="Executing migration" id="add index dashboard_permission" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.85303294Z level=info msg="Migration successfully executed" id="add index dashboard_permission" duration=408.164µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.853458587Z level=info msg="Executing migration" id="save default acl rules in dashboard_acl table" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.853723703Z level=info msg="Migration successfully executed" id="save default acl rules in dashboard_acl table" duration=263.573µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.854246391Z level=info msg="Executing migration" id="delete acl rules for deleted dashboards and folders" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.85435757Z level=info msg="Migration successfully executed" id="delete acl rules for deleted dashboards and folders" duration=111.168µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.854799165Z level=info msg="Executing migration" id="create tag table" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.855141617Z level=info msg="Migration successfully executed" id="create tag table" duration=342.231µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.855622657Z level=info msg="Executing migration" id="add index tag.key_value" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.85600873Z level=info msg="Migration successfully executed" id="add index tag.key_value" duration=386.053µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.856463952Z level=info msg="Executing migration" id="create login attempt table" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.85679381Z level=info msg="Migration successfully executed" id="create login attempt table" duration=329.788µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.857262327Z level=info msg="Executing migration" id="add index login_attempt.username" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.857670791Z level=info msg="Migration successfully executed" id="add index login_attempt.username" duration=408.254µs 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.858133898Z level=info msg="Executing migration" id="drop index IDX_login_attempt_username - v1" 2026-03-10T11:32:39.915 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.858537553Z level=info msg="Migration successfully executed" id="drop index IDX_login_attempt_username - v1" duration=403.695µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.85900602Z level=info msg="Executing migration" id="Rename table login_attempt to login_attempt_tmp_qwerty - v1" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.862921239Z level=info msg="Migration successfully executed" id="Rename table login_attempt to login_attempt_tmp_qwerty - v1" duration=3.914879ms 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.863443456Z level=info msg="Executing migration" id="create login_attempt v2" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.863796707Z level=info msg="Migration successfully executed" id="create login_attempt v2" duration=355.426µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.864216493Z level=info msg="Executing migration" id="create index IDX_login_attempt_username - v2" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.864619438Z level=info msg="Migration successfully executed" id="create index IDX_login_attempt_username - v2" duration=402.685µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.865092883Z level=info msg="Executing migration" id="copy login_attempt v1 to v2" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.865248154Z level=info msg="Migration successfully executed" id="copy login_attempt v1 to v2" duration=155.24µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.865681535Z level=info msg="Executing migration" id="drop login_attempt_tmp_qwerty" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.865958413Z level=info msg="Migration successfully executed" id="drop login_attempt_tmp_qwerty" duration=276.867µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.866393829Z level=info msg="Executing migration" id="create user auth table" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.866714729Z level=info msg="Migration successfully executed" id="create user auth table" duration=320.771µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.867154121Z level=info msg="Executing migration" id="create index IDX_user_auth_auth_module_auth_id - v1" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.867565312Z level=info msg="Migration successfully executed" id="create index IDX_user_auth_auth_module_auth_id - v1" duration=411.061µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.868016385Z level=info msg="Executing migration" id="alter user_auth.auth_id to length 190" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.868043286Z level=info msg="Migration successfully executed" id="alter user_auth.auth_id to length 190" duration=27.341µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.868545796Z level=info msg="Executing migration" id="Add OAuth access token to user_auth" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.870088634Z level=info msg="Migration successfully executed" id="Add OAuth access token to user_auth" duration=1.542707ms 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.870552593Z level=info msg="Executing migration" id="Add OAuth refresh token to user_auth" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.872077967Z level=info msg="Migration successfully executed" id="Add OAuth refresh token to user_auth" duration=1.525124ms 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.872579737Z level=info msg="Executing migration" id="Add OAuth token type to user_auth" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.874105373Z level=info msg="Migration successfully executed" id="Add OAuth token type to user_auth" duration=1.525184ms 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.874630095Z level=info msg="Executing migration" id="Add OAuth expiry to user_auth" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.876167262Z level=info msg="Migration successfully executed" id="Add OAuth expiry to user_auth" duration=1.536537ms 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.876625379Z level=info msg="Executing migration" id="Add index to user_id column in user_auth" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.877073248Z level=info msg="Migration successfully executed" id="Add index to user_id column in user_auth" duration=445.795µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.877591287Z level=info msg="Executing migration" id="Add OAuth ID token to user_auth" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.879332557Z level=info msg="Migration successfully executed" id="Add OAuth ID token to user_auth" duration=1.740979ms 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.879854134Z level=info msg="Executing migration" id="create server_lock table" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.880250424Z level=info msg="Migration successfully executed" id="create server_lock table" duration=396.341µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.880769797Z level=info msg="Executing migration" id="add index server_lock.operation_uid" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.881205442Z level=info msg="Migration successfully executed" id="add index server_lock.operation_uid" duration=433.682µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.881683687Z level=info msg="Executing migration" id="create user auth token table" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.882036968Z level=info msg="Migration successfully executed" id="create user auth token table" duration=352.641µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.882518249Z level=info msg="Executing migration" id="add unique index user_auth_token.auth_token" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.882934659Z level=info msg="Migration successfully executed" id="add unique index user_auth_token.auth_token" duration=416.3µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.883371877Z level=info msg="Executing migration" id="add unique index user_auth_token.prev_auth_token" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.883787275Z level=info msg="Migration successfully executed" id="add unique index user_auth_token.prev_auth_token" duration=415.458µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.88427107Z level=info msg="Executing migration" id="add index user_auth_token.user_id" 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.884718828Z level=info msg="Migration successfully executed" id="add index user_auth_token.user_id" duration=447.569µs 2026-03-10T11:32:39.916 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.885194999Z level=info msg="Executing migration" id="Add revoked_at to the user auth token" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.886930198Z level=info msg="Migration successfully executed" id="Add revoked_at to the user auth token" duration=1.735027ms 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.887495766Z level=info msg="Executing migration" id="add index user_auth_token.revoked_at" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.887997976Z level=info msg="Migration successfully executed" id="add index user_auth_token.revoked_at" duration=501.819µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.888498302Z level=info msg="Executing migration" id="create cache_data table" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.888882412Z level=info msg="Migration successfully executed" id="create cache_data table" duration=384.069µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.889363121Z level=info msg="Executing migration" id="add unique index cache_data.cache_key" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.88977383Z level=info msg="Migration successfully executed" id="add unique index cache_data.cache_key" duration=410.739µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.890237107Z level=info msg="Executing migration" id="create short_url table v1" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.890622939Z level=info msg="Migration successfully executed" id="create short_url table v1" duration=385.773µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.891102607Z level=info msg="Executing migration" id="add index short_url.org_id-uid" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.891587365Z level=info msg="Migration successfully executed" id="add index short_url.org_id-uid" duration=483.526µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.892137073Z level=info msg="Executing migration" id="alter table short_url alter column created_by type to bigint" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.892164284Z level=info msg="Migration successfully executed" id="alter table short_url alter column created_by type to bigint" duration=27.752µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.892700098Z level=info msg="Executing migration" id="delete alert_definition table" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.892742818Z level=info msg="Migration successfully executed" id="delete alert_definition table" duration=42.901µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.893263762Z level=info msg="Executing migration" id="recreate alert_definition table" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.893656628Z level=info msg="Migration successfully executed" id="recreate alert_definition table" duration=392.826µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.894121587Z level=info msg="Executing migration" id="add index in alert_definition on org_id and title columns" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.894537296Z level=info msg="Migration successfully executed" id="add index in alert_definition on org_id and title columns" duration=415.539µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.895027243Z level=info msg="Executing migration" id="add index in alert_definition on org_id and uid columns" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.895439675Z level=info msg="Migration successfully executed" id="add index in alert_definition on org_id and uid columns" duration=412.343µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.895896861Z level=info msg="Executing migration" id="alter alert_definition table data column to mediumtext in mysql" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.89592336Z level=info msg="Migration successfully executed" id="alter alert_definition table data column to mediumtext in mysql" duration=26.85µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.896426553Z level=info msg="Executing migration" id="drop index in alert_definition on org_id and title columns" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.896884449Z level=info msg="Migration successfully executed" id="drop index in alert_definition on org_id and title columns" duration=458.748µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.897308985Z level=info msg="Executing migration" id="drop index in alert_definition on org_id and uid columns" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.897722067Z level=info msg="Migration successfully executed" id="drop index in alert_definition on org_id and uid columns" duration=413.062µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.898165768Z level=info msg="Executing migration" id="add unique index in alert_definition on org_id and title columns" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.898610039Z level=info msg="Migration successfully executed" id="add unique index in alert_definition on org_id and title columns" duration=444.071µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.899073447Z level=info msg="Executing migration" id="add unique index in alert_definition on org_id and uid columns" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.899498091Z level=info msg="Migration successfully executed" id="add unique index in alert_definition on org_id and uid columns" duration=424.525µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.900008336Z level=info msg="Executing migration" id="Add column paused in alert_definition" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.901868017Z level=info msg="Migration successfully executed" id="Add column paused in alert_definition" duration=1.85913ms 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.902451259Z level=info msg="Executing migration" id="drop alert_definition table" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.903034872Z level=info msg="Migration successfully executed" id="drop alert_definition table" duration=583.583µs 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.903618644Z level=info msg="Executing migration" id="delete alert_definition_version table" 2026-03-10T11:32:39.917 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.903694696Z level=info msg="Migration successfully executed" id="delete alert_definition_version table" duration=76.493µs 2026-03-10T11:32:39.918 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.904280834Z level=info msg="Executing migration" id="recreate alert_definition_version table" 2026-03-10T11:32:39.918 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.904800415Z level=info msg="Migration successfully executed" id="recreate alert_definition_version table" duration=519.663µs 2026-03-10T11:32:40.159 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.91278401Z level=info msg="Executing migration" id="add index in alert_definition_version table on alert_definition_id and version columns" 2026-03-10T11:32:40.159 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.913450086Z level=info msg="Migration successfully executed" id="add index in alert_definition_version table on alert_definition_id and version columns" duration=667.96µs 2026-03-10T11:32:40.159 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.914044719Z level=info msg="Executing migration" id="add index in alert_definition_version table on alert_definition_uid and version columns" 2026-03-10T11:32:40.159 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.91452044Z level=info msg="Migration successfully executed" id="add index in alert_definition_version table on alert_definition_uid and version columns" duration=476.422µs 2026-03-10T11:32:40.159 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.915029522Z level=info msg="Executing migration" id="alter alert_definition_version table data column to mediumtext in mysql" 2026-03-10T11:32:40.159 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.915067253Z level=info msg="Migration successfully executed" id="alter alert_definition_version table data column to mediumtext in mysql" duration=37.971µs 2026-03-10T11:32:40.159 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.915617222Z level=info msg="Executing migration" id="drop alert_definition_version table" 2026-03-10T11:32:40.159 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.916117068Z level=info msg="Migration successfully executed" id="drop alert_definition_version table" duration=499.705µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.916724185Z level=info msg="Executing migration" id="create alert_instance table" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.917175398Z level=info msg="Migration successfully executed" id="create alert_instance table" duration=451.044µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.917698939Z level=info msg="Executing migration" id="add index in alert_instance table on def_org_id, def_uid and current_state columns" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.918148009Z level=info msg="Migration successfully executed" id="add index in alert_instance table on def_org_id, def_uid and current_state columns" duration=448.809µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.918612399Z level=info msg="Executing migration" id="add index in alert_instance table on def_org_id, current_state columns" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.919022406Z level=info msg="Migration successfully executed" id="add index in alert_instance table on def_org_id, current_state columns" duration=430.014µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.919490563Z level=info msg="Executing migration" id="add column current_state_end to alert_instance" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.921272668Z level=info msg="Migration successfully executed" id="add column current_state_end to alert_instance" duration=1.781875ms 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.92176034Z level=info msg="Executing migration" id="remove index def_org_id, def_uid, current_state on alert_instance" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.922179546Z level=info msg="Migration successfully executed" id="remove index def_org_id, def_uid, current_state on alert_instance" duration=419.045µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.922665846Z level=info msg="Executing migration" id="remove index def_org_id, current_state on alert_instance" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.923067507Z level=info msg="Migration successfully executed" id="remove index def_org_id, current_state on alert_instance" duration=401.531µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.923550231Z level=info msg="Executing migration" id="rename def_org_id to rule_org_id in alert_instance" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.931924886Z level=info msg="Migration successfully executed" id="rename def_org_id to rule_org_id in alert_instance" duration=8.371319ms 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.932636588Z level=info msg="Executing migration" id="rename def_uid to rule_uid in alert_instance" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.940402183Z level=info msg="Migration successfully executed" id="rename def_uid to rule_uid in alert_instance" duration=7.761347ms 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.941261292Z level=info msg="Executing migration" id="add index rule_org_id, rule_uid, current_state on alert_instance" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.941857067Z level=info msg="Migration successfully executed" id="add index rule_org_id, rule_uid, current_state on alert_instance" duration=595.555µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.942391889Z level=info msg="Executing migration" id="add index rule_org_id, current_state on alert_instance" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.942879621Z level=info msg="Migration successfully executed" id="add index rule_org_id, current_state on alert_instance" duration=485.267µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.943397351Z level=info msg="Executing migration" id="add current_reason column related to current_state" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.945301865Z level=info msg="Migration successfully executed" id="add current_reason column related to current_state" duration=1.904095ms 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.94583378Z level=info msg="Executing migration" id="add result_fingerprint column to alert_instance" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.947497675Z level=info msg="Migration successfully executed" id="add result_fingerprint column to alert_instance" duration=1.66147ms 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.948034028Z level=info msg="Executing migration" id="create alert_rule table" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.948439658Z level=info msg="Migration successfully executed" id="create alert_rule table" duration=405.75µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.94898063Z level=info msg="Executing migration" id="add index in alert_rule on org_id and title columns" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.949421185Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id and title columns" duration=439.003µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.949912344Z level=info msg="Executing migration" id="add index in alert_rule on org_id and uid columns" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.950341508Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id and uid columns" duration=428.714µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.950820133Z level=info msg="Executing migration" id="add index in alert_rule on org_id, namespace_uid, group_uid columns" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.951289563Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, namespace_uid, group_uid columns" duration=469.108µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.951832227Z level=info msg="Executing migration" id="alter alert_rule table data column to mediumtext in mysql" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.951864638Z level=info msg="Migration successfully executed" id="alter alert_rule table data column to mediumtext in mysql" duration=32.551µs 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.952381676Z level=info msg="Executing migration" id="add column for to alert_rule" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.954267577Z level=info msg="Migration successfully executed" id="add column for to alert_rule" duration=1.885459ms 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.954822965Z level=info msg="Executing migration" id="add column annotations to alert_rule" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.956810817Z level=info msg="Migration successfully executed" id="add column annotations to alert_rule" duration=1.987301ms 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.957364914Z level=info msg="Executing migration" id="add column labels to alert_rule" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.959259149Z level=info msg="Migration successfully executed" id="add column labels to alert_rule" duration=1.892091ms 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.959771629Z level=info msg="Executing migration" id="remove unique index from alert_rule on org_id, title columns" 2026-03-10T11:32:40.160 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.960204709Z level=info msg="Migration successfully executed" id="remove unique index from alert_rule on org_id, title columns" duration=432.83µs 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.960691561Z level=info msg="Executing migration" id="add index in alert_rule on org_id, namespase_uid and title columns" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.961125252Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, namespase_uid and title columns" duration=433.661µs 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.961586406Z level=info msg="Executing migration" id="add dashboard_uid column to alert_rule" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.963274004Z level=info msg="Migration successfully executed" id="add dashboard_uid column to alert_rule" duration=1.687337ms 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.963826578Z level=info msg="Executing migration" id="add panel_id column to alert_rule" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.965514598Z level=info msg="Migration successfully executed" id="add panel_id column to alert_rule" duration=1.687248ms 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.966035663Z level=info msg="Executing migration" id="add index in alert_rule on org_id, dashboard_uid and panel_id columns" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.966459587Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, dashboard_uid and panel_id columns" duration=424.025µs 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.966990471Z level=info msg="Executing migration" id="add rule_group_idx column to alert_rule" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.968696324Z level=info msg="Migration successfully executed" id="add rule_group_idx column to alert_rule" duration=1.705853ms 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.969204384Z level=info msg="Executing migration" id="add is_paused column to alert_rule table" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.97098704Z level=info msg="Migration successfully executed" id="add is_paused column to alert_rule table" duration=1.780062ms 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.971483751Z level=info msg="Executing migration" id="fix is_paused column for alert_rule table" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.971509819Z level=info msg="Migration successfully executed" id="fix is_paused column for alert_rule table" duration=26.429µs 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.972069206Z level=info msg="Executing migration" id="create alert_rule_version table" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.972542543Z level=info msg="Migration successfully executed" id="create alert_rule_version table" duration=471.493µs 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.973075008Z level=info msg="Executing migration" id="add index in alert_rule_version table on rule_org_id, rule_uid and version columns" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.973547353Z level=info msg="Migration successfully executed" id="add index in alert_rule_version table on rule_org_id, rule_uid and version columns" duration=471.693µs 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.974040697Z level=info msg="Executing migration" id="add index in alert_rule_version table on rule_org_id, rule_namespace_uid and rule_group columns" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.974511268Z level=info msg="Migration successfully executed" id="add index in alert_rule_version table on rule_org_id, rule_namespace_uid and rule_group columns" duration=471.903µs 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.975011915Z level=info msg="Executing migration" id="alter alert_rule_version table data column to mediumtext in mysql" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.975039406Z level=info msg="Migration successfully executed" id="alter alert_rule_version table data column to mediumtext in mysql" duration=27.602µs 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.97554389Z level=info msg="Executing migration" id="add column for to alert_rule_version" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.97754689Z level=info msg="Migration successfully executed" id="add column for to alert_rule_version" duration=2.00288ms 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.978198479Z level=info msg="Executing migration" id="add column annotations to alert_rule_version" 2026-03-10T11:32:40.161 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.980845985Z level=info msg="Migration successfully executed" id="add column annotations to alert_rule_version" duration=2.647355ms 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.981449775Z level=info msg="Executing migration" id="add column labels to alert_rule_version" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.984033781Z level=info msg="Migration successfully executed" id="add column labels to alert_rule_version" duration=2.583426ms 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.984630589Z level=info msg="Executing migration" id="add rule_group_idx column to alert_rule_version" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.987197052Z level=info msg="Migration successfully executed" id="add rule_group_idx column to alert_rule_version" duration=2.568778ms 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.987884299Z level=info msg="Executing migration" id="add is_paused column to alert_rule_versions table" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.990460261Z level=info msg="Migration successfully executed" id="add is_paused column to alert_rule_versions table" duration=2.575372ms 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.991090008Z level=info msg="Executing migration" id="fix is_paused column for alert_rule_version table" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.991173335Z level=info msg="Migration successfully executed" id="fix is_paused column for alert_rule_version table" duration=83.647µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.99177436Z level=info msg="Executing migration" id=create_alert_configuration_table 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.99226069Z level=info msg="Migration successfully executed" id=create_alert_configuration_table duration=486.1µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.99281627Z level=info msg="Executing migration" id="Add column default in alert_configuration" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.995569223Z level=info msg="Migration successfully executed" id="Add column default in alert_configuration" duration=2.75126ms 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.996323525Z level=info msg="Executing migration" id="alert alert_configuration alertmanager_configuration column from TEXT to MEDIUMTEXT if mysql" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.996419625Z level=info msg="Migration successfully executed" id="alert alert_configuration alertmanager_configuration column from TEXT to MEDIUMTEXT if mysql" duration=96.22µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.997244529Z level=info msg="Executing migration" id="add column org_id in alert_configuration" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:39.999814288Z level=info msg="Migration successfully executed" id="add column org_id in alert_configuration" duration=2.567785ms 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.000525369Z level=info msg="Executing migration" id="add index in alert_configuration table on org_id column" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.001079106Z level=info msg="Migration successfully executed" id="add index in alert_configuration table on org_id column" duration=551.974µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.00169038Z level=info msg="Executing migration" id="add configuration_hash column to alert_configuration" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.003569347Z level=info msg="Migration successfully executed" id="add configuration_hash column to alert_configuration" duration=1.878766ms 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.004164641Z level=info msg="Executing migration" id=create_ngalert_configuration_table 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.004535375Z level=info msg="Migration successfully executed" id=create_ngalert_configuration_table duration=370.624µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.005076058Z level=info msg="Executing migration" id="add index in ngalert_configuration on org_id column" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.005502856Z level=info msg="Migration successfully executed" id="add index in ngalert_configuration on org_id column" duration=426.789µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.006050221Z level=info msg="Executing migration" id="add column send_alerts_to in ngalert_configuration" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.007957832Z level=info msg="Migration successfully executed" id="add column send_alerts_to in ngalert_configuration" duration=1.906889ms 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.008498533Z level=info msg="Executing migration" id="create provenance_type table" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.008888754Z level=info msg="Migration successfully executed" id="create provenance_type table" duration=390.071µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.009433182Z level=info msg="Executing migration" id="add index to uniquify (record_key, record_type, org_id) columns" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.009917369Z level=info msg="Migration successfully executed" id="add index to uniquify (record_key, record_type, org_id) columns" duration=484.948µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.010408868Z level=info msg="Executing migration" id="create alert_image table" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.010790744Z level=info msg="Migration successfully executed" id="create alert_image table" duration=381.886µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.017332619Z level=info msg="Executing migration" id="add unique index on token to alert_image table" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.018274953Z level=info msg="Migration successfully executed" id="add unique index on token to alert_image table" duration=944.258µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.019009878Z level=info msg="Executing migration" id="support longer URLs in alert_image table" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.019101058Z level=info msg="Migration successfully executed" id="support longer URLs in alert_image table" duration=91.051µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.019810286Z level=info msg="Executing migration" id=create_alert_configuration_history_table 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.020454362Z level=info msg="Migration successfully executed" id=create_alert_configuration_history_table duration=643.816µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.021159271Z level=info msg="Executing migration" id="drop non-unique orgID index on alert_configuration" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.021776446Z level=info msg="Migration successfully executed" id="drop non-unique orgID index on alert_configuration" duration=617.024µs 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.022341925Z level=info msg="Executing migration" id="drop unique orgID index on alert_configuration if exists" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.022615316Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop unique orgID index on alert_configuration if exists" 2026-03-10T11:32:40.162 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.023244144Z level=info msg="Executing migration" id="extract alertmanager configuration history to separate table" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.023527234Z level=info msg="Migration successfully executed" id="extract alertmanager configuration history to separate table" duration=282.979µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.024069669Z level=info msg="Executing migration" id="add unique index on orgID to alert_configuration" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.02469011Z level=info msg="Migration successfully executed" id="add unique index on orgID to alert_configuration" duration=620.29µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.0252404Z level=info msg="Executing migration" id="add last_applied column to alert_configuration_history" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.028018861Z level=info msg="Migration successfully executed" id="add last_applied column to alert_configuration_history" duration=2.778391ms 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.02861165Z level=info msg="Executing migration" id="create library_element table v1" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.029128407Z level=info msg="Migration successfully executed" id="create library_element table v1" duration=516.467µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.029697291Z level=info msg="Executing migration" id="add index library_element org_id-folder_id-name-kind" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.030166501Z level=info msg="Migration successfully executed" id="add index library_element org_id-folder_id-name-kind" duration=468.718µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.030721069Z level=info msg="Executing migration" id="create library_element_connection table v1" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.03111171Z level=info msg="Migration successfully executed" id="create library_element_connection table v1" duration=389.328µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.031655777Z level=info msg="Executing migration" id="add index library_element_connection element_id-kind-connection_id" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.032199014Z level=info msg="Migration successfully executed" id="add index library_element_connection element_id-kind-connection_id" duration=542.957µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.032762389Z level=info msg="Executing migration" id="add unique index library_element org_id_uid" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.033205729Z level=info msg="Migration successfully executed" id="add unique index library_element org_id_uid" duration=444.312µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.033694525Z level=info msg="Executing migration" id="increase max description length to 2048" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.033704263Z level=info msg="Migration successfully executed" id="increase max description length to 2048" duration=10.36µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.034254973Z level=info msg="Executing migration" id="alter library_element model to mediumtext" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.034330384Z level=info msg="Migration successfully executed" id="alter library_element model to mediumtext" duration=99.958µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.034729912Z level=info msg="Executing migration" id="clone move dashboard alerts to unified alerting" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.034906382Z level=info msg="Migration successfully executed" id="clone move dashboard alerts to unified alerting" duration=176.421µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.035429772Z level=info msg="Executing migration" id="create data_keys table" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.035883241Z level=info msg="Migration successfully executed" id="create data_keys table" duration=453.42µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.036407502Z level=info msg="Executing migration" id="create secrets table" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.036840874Z level=info msg="Migration successfully executed" id="create secrets table" duration=433.191µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.037310712Z level=info msg="Executing migration" id="rename data_keys name column to id" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.047961588Z level=info msg="Migration successfully executed" id="rename data_keys name column to id" duration=10.637982ms 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.048755885Z level=info msg="Executing migration" id="add name column into data_keys" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.051589629Z level=info msg="Migration successfully executed" id="add name column into data_keys" duration=2.833904ms 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.052164124Z level=info msg="Executing migration" id="copy data_keys id column values into name" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.052283868Z level=info msg="Migration successfully executed" id="copy data_keys id column values into name" duration=119.714µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.052827377Z level=info msg="Executing migration" id="rename data_keys name column to label" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.062705987Z level=info msg="Migration successfully executed" id="rename data_keys name column to label" duration=9.877688ms 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.065691675Z level=info msg="Executing migration" id="rename data_keys id column back to name" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.080840059Z level=info msg="Migration successfully executed" id="rename data_keys id column back to name" duration=15.14576ms 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.081673198Z level=info msg="Executing migration" id="create kv_store table v1" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.082106079Z level=info msg="Migration successfully executed" id="create kv_store table v1" duration=432.8µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.082636882Z level=info msg="Executing migration" id="add index kv_store.org_id-namespace-key" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.08317028Z level=info msg="Migration successfully executed" id="add index kv_store.org_id-namespace-key" duration=533.017µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.083797705Z level=info msg="Executing migration" id="update dashboard_uid and panel_id from existing annotations" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.083960029Z level=info msg="Migration successfully executed" id="update dashboard_uid and panel_id from existing annotations" duration=162.424µs 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.084611809Z level=info msg="Executing migration" id="create permission table" 2026-03-10T11:32:40.163 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.085082189Z level=info msg="Migration successfully executed" id="create permission table" duration=489.445µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.085664119Z level=info msg="Executing migration" id="add unique index permission.role_id" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.086161048Z level=info msg="Migration successfully executed" id="add unique index permission.role_id" duration=496.99µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.086693535Z level=info msg="Executing migration" id="add unique index role_id_action_scope" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.08723584Z level=info msg="Migration successfully executed" id="add unique index role_id_action_scope" duration=540.571µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.087785379Z level=info msg="Executing migration" id="create role table" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.088194484Z level=info msg="Migration successfully executed" id="create role table" duration=408.865µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.088701193Z level=info msg="Executing migration" id="add column display_name" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.09092214Z level=info msg="Migration successfully executed" id="add column display_name" duration=2.221207ms 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.091428568Z level=info msg="Executing migration" id="add column group_name" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.093563965Z level=info msg="Migration successfully executed" id="add column group_name" duration=2.134745ms 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.094168667Z level=info msg="Executing migration" id="add index role.org_id" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.094656209Z level=info msg="Migration successfully executed" id="add index role.org_id" duration=487.552µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.095164381Z level=info msg="Executing migration" id="add unique index role_org_id_name" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.095692188Z level=info msg="Migration successfully executed" id="add unique index role_org_id_name" duration=527.758µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.096184731Z level=info msg="Executing migration" id="add index role_org_id_uid" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.096668726Z level=info msg="Migration successfully executed" id="add index role_org_id_uid" duration=483.965µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.097186025Z level=info msg="Executing migration" id="create team role table" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.09758474Z level=info msg="Migration successfully executed" id="create team role table" duration=397.434µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.098113541Z level=info msg="Executing migration" id="add index team_role.org_id" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.098668168Z level=info msg="Migration successfully executed" id="add index team_role.org_id" duration=554.688µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.099210143Z level=info msg="Executing migration" id="add unique index team_role_org_id_team_id_role_id" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.099798033Z level=info msg="Migration successfully executed" id="add unique index team_role_org_id_team_id_role_id" duration=587.891µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.100374092Z level=info msg="Executing migration" id="add index team_role.team_id" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.100881371Z level=info msg="Migration successfully executed" id="add index team_role.team_id" duration=507.36µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.101378381Z level=info msg="Executing migration" id="create user role table" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.101825798Z level=info msg="Migration successfully executed" id="create user role table" duration=447.248µs 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.102349579Z level=info msg="Executing migration" id="add index user_role.org_id" 2026-03-10T11:32:40.164 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.102822816Z level=info msg="Migration successfully executed" id="add index user_role.org_id" duration=473.267µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.103327369Z level=info msg="Executing migration" id="add unique index user_role_org_id_user_id_role_id" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.10382443Z level=info msg="Migration successfully executed" id="add unique index user_role_org_id_user_id_role_id" duration=497.03µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.104342569Z level=info msg="Executing migration" id="add index user_role.user_id" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.10479698Z level=info msg="Migration successfully executed" id="add index user_role.user_id" duration=454.361µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.10537359Z level=info msg="Executing migration" id="create builtin role table" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.10575837Z level=info msg="Migration successfully executed" id="create builtin role table" duration=384.881µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.10625007Z level=info msg="Executing migration" id="add index builtin_role.role_id" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.106733225Z level=info msg="Migration successfully executed" id="add index builtin_role.role_id" duration=483.094µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.10723873Z level=info msg="Executing migration" id="add index builtin_role.name" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.107735981Z level=info msg="Migration successfully executed" id="add index builtin_role.name" duration=497.211µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.108224947Z level=info msg="Executing migration" id="Add column org_id to builtin_role table" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.110684079Z level=info msg="Migration successfully executed" id="Add column org_id to builtin_role table" duration=2.45747ms 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.111173695Z level=info msg="Executing migration" id="add index builtin_role.org_id" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.111655678Z level=info msg="Migration successfully executed" id="add index builtin_role.org_id" duration=481.903µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.112135486Z level=info msg="Executing migration" id="add unique index builtin_role_org_id_role_id_role" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.112620775Z level=info msg="Migration successfully executed" id="add unique index builtin_role_org_id_role_id_role" duration=485.168µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.11309935Z level=info msg="Executing migration" id="Remove unique index role_org_id_uid" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.113575362Z level=info msg="Migration successfully executed" id="Remove unique index role_org_id_uid" duration=476.021µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.114063745Z level=info msg="Executing migration" id="add unique index role.uid" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.114527173Z level=info msg="Migration successfully executed" id="add unique index role.uid" duration=463.438µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.115061011Z level=info msg="Executing migration" id="create seed assignment table" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.115415295Z level=info msg="Migration successfully executed" id="create seed assignment table" duration=352.891µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.11591519Z level=info msg="Executing migration" id="add unique index builtin_role_role_name" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.11637929Z level=info msg="Migration successfully executed" id="add unique index builtin_role_role_name" duration=464.039µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.117066525Z level=info msg="Executing migration" id="add column hidden to role table" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.1194726Z level=info msg="Migration successfully executed" id="add column hidden to role table" duration=2.404861ms 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.120008933Z level=info msg="Executing migration" id="permission kind migration" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.122352139Z level=info msg="Migration successfully executed" id="permission kind migration" duration=2.343006ms 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.12284437Z level=info msg="Executing migration" id="permission attribute migration" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.125203857Z level=info msg="Migration successfully executed" id="permission attribute migration" duration=2.359246ms 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.125758054Z level=info msg="Executing migration" id="permission identifier migration" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.128326401Z level=info msg="Migration successfully executed" id="permission identifier migration" duration=2.568107ms 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.128818403Z level=info msg="Executing migration" id="add permission identifier index" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.129348574Z level=info msg="Migration successfully executed" id="add permission identifier index" duration=529.991µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.129836488Z level=info msg="Executing migration" id="add permission action scope role_id index" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.13035018Z level=info msg="Migration successfully executed" id="add permission action scope role_id index" duration=513.63µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.130860004Z level=info msg="Executing migration" id="remove permission role_id action scope index" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.131326657Z level=info msg="Migration successfully executed" id="remove permission role_id action scope index" duration=466.392µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.131856409Z level=info msg="Executing migration" id="create query_history table v1" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.132259473Z level=info msg="Migration successfully executed" id="create query_history table v1" duration=403.384µs 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.132750513Z level=info msg="Executing migration" id="add index query_history.org_id-created_by-datasource_uid" 2026-03-10T11:32:40.165 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.133212377Z level=info msg="Migration successfully executed" id="add index query_history.org_id-created_by-datasource_uid" duration=462.896µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.133751035Z level=info msg="Executing migration" id="alter table query_history alter column created_by type to bigint" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.133808393Z level=info msg="Migration successfully executed" id="alter table query_history alter column created_by type to bigint" duration=57.948µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.134308578Z level=info msg="Executing migration" id="rbac disabled migrator" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.134347201Z level=info msg="Migration successfully executed" id="rbac disabled migrator" duration=39.013µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.134752189Z level=info msg="Executing migration" id="teams permissions migration" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.13498774Z level=info msg="Migration successfully executed" id="teams permissions migration" duration=235.751µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.13550615Z level=info msg="Executing migration" id="dashboard permissions" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.135802094Z level=info msg="Migration successfully executed" id="dashboard permissions" duration=297.517µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.136308352Z level=info msg="Executing migration" id="dashboard permissions uid scopes" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.136613303Z level=info msg="Migration successfully executed" id="dashboard permissions uid scopes" duration=284.382µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.137092419Z level=info msg="Executing migration" id="drop managed folder create actions" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.137218425Z level=info msg="Migration successfully executed" id="drop managed folder create actions" duration=125.906µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.137694296Z level=info msg="Executing migration" id="alerting notification permissions" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.137957539Z level=info msg="Migration successfully executed" id="alerting notification permissions" duration=263.262µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.138450491Z level=info msg="Executing migration" id="create query_history_star table v1" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.138836153Z level=info msg="Migration successfully executed" id="create query_history_star table v1" duration=385.671µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.139345636Z level=info msg="Executing migration" id="add index query_history.user_id-query_uid" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.139850612Z level=info msg="Migration successfully executed" id="add index query_history.user_id-query_uid" duration=505.046µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.140329769Z level=info msg="Executing migration" id="add column org_id in query_history_star" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.142806435Z level=info msg="Migration successfully executed" id="add column org_id in query_history_star" duration=2.475293ms 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.143324986Z level=info msg="Executing migration" id="alter table query_history_star_mig column user_id type to bigint" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.143383575Z level=info msg="Migration successfully executed" id="alter table query_history_star_mig column user_id type to bigint" duration=59.021µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.143916341Z level=info msg="Executing migration" id="create correlation table v1" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.144410517Z level=info msg="Migration successfully executed" id="create correlation table v1" duration=471.723µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.1449464Z level=info msg="Executing migration" id="add index correlations.uid" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.145409006Z level=info msg="Migration successfully executed" id="add index correlations.uid" duration=462.676µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.145888553Z level=info msg="Executing migration" id="add index correlations.source_uid" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.146361499Z level=info msg="Migration successfully executed" id="add index correlations.source_uid" duration=473.005µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.146835616Z level=info msg="Executing migration" id="add correlation config column" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.149222534Z level=info msg="Migration successfully executed" id="add correlation config column" duration=2.386628ms 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.149734362Z level=info msg="Executing migration" id="drop index IDX_correlation_uid - v1" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.150213158Z level=info msg="Migration successfully executed" id="drop index IDX_correlation_uid - v1" duration=477.535µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.150702664Z level=info msg="Executing migration" id="drop index IDX_correlation_source_uid - v1" 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.151162976Z level=info msg="Migration successfully executed" id="drop index IDX_correlation_source_uid - v1" duration=461.544µs 2026-03-10T11:32:40.166 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.15168876Z level=info msg="Executing migration" id="Rename table correlation to correlation_tmp_qwerty - v1" 2026-03-10T11:32:40.167 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:40 vm09 ceph-mon[54793]: pgmap v10: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 0 B/s wr, 6 op/s 2026-03-10T11:32:40.167 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:40 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.167 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:40 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.167 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:40 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.167 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:40 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.15801423Z level=info msg="Migration successfully executed" id="Rename table correlation to correlation_tmp_qwerty - v1" duration=6.324318ms 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.158635873Z level=info msg="Executing migration" id="create correlation v2" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.1591524Z level=info msg="Migration successfully executed" id="create correlation v2" duration=516.336µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.163871793Z level=info msg="Executing migration" id="create index IDX_correlation_uid - v2" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.164515158Z level=info msg="Migration successfully executed" id="create index IDX_correlation_uid - v2" duration=644.507µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.165210469Z level=info msg="Executing migration" id="create index IDX_correlation_source_uid - v2" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.16580425Z level=info msg="Migration successfully executed" id="create index IDX_correlation_source_uid - v2" duration=593.802µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.166385569Z level=info msg="Executing migration" id="create index IDX_correlation_org_id - v2" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.166932302Z level=info msg="Migration successfully executed" id="create index IDX_correlation_org_id - v2" duration=546.692µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.167639425Z level=info msg="Executing migration" id="copy correlation v1 to v2" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.167826636Z level=info msg="Migration successfully executed" id="copy correlation v1 to v2" duration=187.54µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.1683785Z level=info msg="Executing migration" id="drop correlation_tmp_qwerty" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.168871342Z level=info msg="Migration successfully executed" id="drop correlation_tmp_qwerty" duration=492.572µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.169467297Z level=info msg="Executing migration" id="add provisioning column" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.172058818Z level=info msg="Migration successfully executed" id="add provisioning column" duration=2.591321ms 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.172655334Z level=info msg="Executing migration" id="create entity_events table" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.173078727Z level=info msg="Migration successfully executed" id="create entity_events table" duration=424.986µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.173645217Z level=info msg="Executing migration" id="create dashboard public config v1" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.174137479Z level=info msg="Migration successfully executed" id="create dashboard public config v1" duration=492.462µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.174681005Z level=info msg="Executing migration" id="drop index UQE_dashboard_public_config_uid - v1" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.174929741Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop index UQE_dashboard_public_config_uid - v1" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.175518253Z level=info msg="Executing migration" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v1" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.175783088Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v1" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.17636089Z level=info msg="Executing migration" id="Drop old dashboard public config table" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.176826552Z level=info msg="Migration successfully executed" id="Drop old dashboard public config table" duration=464.229µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.177316208Z level=info msg="Executing migration" id="recreate dashboard public config v1" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.177821164Z level=info msg="Migration successfully executed" id="recreate dashboard public config v1" duration=503.443µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.1783536Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_uid - v1" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.178885436Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_uid - v1" duration=531.675µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.179418112Z level=info msg="Executing migration" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v1" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.179991566Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v1" duration=573.624µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.180546334Z level=info msg="Executing migration" id="drop index UQE_dashboard_public_config_uid - v2" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.18107816Z level=info msg="Migration successfully executed" id="drop index UQE_dashboard_public_config_uid - v2" duration=530.413µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.181622799Z level=info msg="Executing migration" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v2" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.18213078Z level=info msg="Migration successfully executed" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v2" duration=508.121µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.182741293Z level=info msg="Executing migration" id="Drop public config table" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.183181046Z level=info msg="Migration successfully executed" id="Drop public config table" duration=439.622µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.183758597Z level=info msg="Executing migration" id="Recreate dashboard public config v2" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.184264243Z level=info msg="Migration successfully executed" id="Recreate dashboard public config v2" duration=505.526µs 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.18478024Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_uid - v2" 2026-03-10T11:32:40.281 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.185290694Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_uid - v2" duration=510.766µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.185803023Z level=info msg="Executing migration" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v2" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.186337102Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v2" duration=534.08µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.186868588Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_access_token - v2" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.187384663Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_access_token - v2" duration=515.985µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.18792768Z level=info msg="Executing migration" id="Rename table dashboard_public_config to dashboard_public - v2" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.19610123Z level=info msg="Migration successfully executed" id="Rename table dashboard_public_config to dashboard_public - v2" duration=8.170584ms 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.197039214Z level=info msg="Executing migration" id="add annotations_enabled column" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.199961725Z level=info msg="Migration successfully executed" id="add annotations_enabled column" duration=2.920087ms 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.200620047Z level=info msg="Executing migration" id="add time_selection_enabled column" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.203015842Z level=info msg="Migration successfully executed" id="add time_selection_enabled column" duration=2.395754ms 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.203658735Z level=info msg="Executing migration" id="delete orphaned public dashboards" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.203834033Z level=info msg="Migration successfully executed" id="delete orphaned public dashboards" duration=175.99µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.204412395Z level=info msg="Executing migration" id="add share column" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.206825613Z level=info msg="Migration successfully executed" id="add share column" duration=2.413378ms 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.207377315Z level=info msg="Executing migration" id="backfill empty share column fields with default of public" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.207545952Z level=info msg="Migration successfully executed" id="backfill empty share column fields with default of public" duration=168.767µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.208172243Z level=info msg="Executing migration" id="create file table" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.20863013Z level=info msg="Migration successfully executed" id="create file table" duration=457.907µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.209170411Z level=info msg="Executing migration" id="file table idx: path natural pk" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.209694533Z level=info msg="Migration successfully executed" id="file table idx: path natural pk" duration=524.763µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.210230155Z level=info msg="Executing migration" id="file table idx: parent_folder_path_hash fast folder retrieval" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.210744387Z level=info msg="Migration successfully executed" id="file table idx: parent_folder_path_hash fast folder retrieval" duration=512.35µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.211272806Z level=info msg="Executing migration" id="create file_meta table" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.21170159Z level=info msg="Migration successfully executed" id="create file_meta table" duration=428.823µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.212221812Z level=info msg="Executing migration" id="file table idx: path key" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.212737197Z level=info msg="Migration successfully executed" id="file table idx: path key" duration=515.214µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.213259995Z level=info msg="Executing migration" id="set path collation in file table" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.213353642Z level=info msg="Migration successfully executed" id="set path collation in file table" duration=94.006µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.213915803Z level=info msg="Executing migration" id="migrate contents column to mediumblob for MySQL" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.214010571Z level=info msg="Migration successfully executed" id="migrate contents column to mediumblob for MySQL" duration=94.967µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.214547987Z level=info msg="Executing migration" id="managed permissions migration" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.214867615Z level=info msg="Migration successfully executed" id="managed permissions migration" duration=319.858µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.215432923Z level=info msg="Executing migration" id="managed folder permissions alert actions migration" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.215618711Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions migration" duration=185.917µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.216205639Z level=info msg="Executing migration" id="RBAC action name migrator" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.216926449Z level=info msg="Migration successfully executed" id="RBAC action name migrator" duration=721.15µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.217521904Z level=info msg="Executing migration" id="Add UID column to playlist" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.220188545Z level=info msg="Migration successfully executed" id="Add UID column to playlist" duration=2.666421ms 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.2207472Z level=info msg="Executing migration" id="Update uid column values in playlist" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.220892111Z level=info msg="Migration successfully executed" id="Update uid column values in playlist" duration=145.072µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.221397227Z level=info msg="Executing migration" id="Add index for uid in playlist" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.221972854Z level=info msg="Migration successfully executed" id="Add index for uid in playlist" duration=575.476µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.222672133Z level=info msg="Executing migration" id="update group index for alert rules" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.222909908Z level=info msg="Migration successfully executed" id="update group index for alert rules" duration=237.254µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.223476128Z level=info msg="Executing migration" id="managed folder permissions alert actions repeated migration" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.223706108Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions repeated migration" duration=230.18µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.224280594Z level=info msg="Executing migration" id="admin only folder/dashboard permission" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.224548175Z level=info msg="Migration successfully executed" id="admin only folder/dashboard permission" duration=266.468µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.225128622Z level=info msg="Executing migration" id="add action column to seed_assignment" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.227645782Z level=info msg="Migration successfully executed" id="add action column to seed_assignment" duration=2.516661ms 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.228186655Z level=info msg="Executing migration" id="add scope column to seed_assignment" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.231738864Z level=info msg="Migration successfully executed" id="add scope column to seed_assignment" duration=3.552349ms 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.232326774Z level=info msg="Executing migration" id="remove unique index builtin_role_role_name before nullable update" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.233009141Z level=info msg="Migration successfully executed" id="remove unique index builtin_role_role_name before nullable update" duration=680.514µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.233572225Z level=info msg="Executing migration" id="update seed_assignment role_name column to nullable" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.2590746Z level=info msg="Migration successfully executed" id="update seed_assignment role_name column to nullable" duration=25.499199ms 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.26004196Z level=info msg="Executing migration" id="add unique index builtin_role_name back" 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.260701415Z level=info msg="Migration successfully executed" id="add unique index builtin_role_name back" duration=659.886µs 2026-03-10T11:32:40.282 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.261182254Z level=info msg="Executing migration" id="add unique index builtin_role_action_scope" 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.261718208Z level=info msg="Migration successfully executed" id="add unique index builtin_role_action_scope" duration=535.844µs 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.262216521Z level=info msg="Executing migration" id="add primary key to seed_assigment" 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.270455362Z level=info msg="Migration successfully executed" id="add primary key to seed_assigment" duration=8.236888ms 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.271362089Z level=info msg="Executing migration" id="add origin column to seed_assignment" 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.273926529Z level=info msg="Migration successfully executed" id="add origin column to seed_assignment" duration=2.56443ms 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.274521693Z level=info msg="Executing migration" id="add origin to plugin seed_assignment" 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.274720304Z level=info msg="Migration successfully executed" id="add origin to plugin seed_assignment" duration=198.921µs 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.275304639Z level=info msg="Executing migration" id="prevent seeding OnCall access" 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.27548734Z level=info msg="Migration successfully executed" id="prevent seeding OnCall access" duration=168.907µs 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.276038873Z level=info msg="Executing migration" id="managed folder permissions alert actions repeated fixed migration" 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.27620852Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions repeated fixed migration" duration=170.029µs 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.276789588Z level=info msg="Executing migration" id="managed folder permissions library panel actions migration" 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.276954416Z level=info msg="Migration successfully executed" id="managed folder permissions library panel actions migration" duration=163.395µs 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.277486432Z level=info msg="Executing migration" id="migrate external alertmanagers to datsourcse" 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.277664595Z level=info msg="Migration successfully executed" id="migrate external alertmanagers to datsourcse" duration=178.583µs 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.278221267Z level=info msg="Executing migration" id="create folder table" 2026-03-10T11:32:40.283 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.278689944Z level=info msg="Migration successfully executed" id="create folder table" duration=468.566µs 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.279222352Z level=info msg="Executing migration" id="Add index for parent_uid" 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.283719659Z level=info msg="Migration successfully executed" id="Add index for parent_uid" duration=4.497147ms 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.284385485Z level=info msg="Executing migration" id="Add unique index for folder.uid and folder.org_id" 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.284950182Z level=info msg="Migration successfully executed" id="Add unique index for folder.uid and folder.org_id" duration=565.528µs 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.285527062Z level=info msg="Executing migration" id="Update folder title length" 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.285646216Z level=info msg="Migration successfully executed" id="Update folder title length" duration=119.423µs 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.286271626Z level=info msg="Executing migration" id="Add unique index for folder.title and folder.parent_uid" 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.286845721Z level=info msg="Migration successfully executed" id="Add unique index for folder.title and folder.parent_uid" duration=574.186µs 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.287402363Z level=info msg="Executing migration" id="Remove unique index for folder.title and folder.parent_uid" 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.288292217Z level=info msg="Migration successfully executed" id="Remove unique index for folder.title and folder.parent_uid" duration=889.394µs 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.288887703Z level=info msg="Executing migration" id="Add unique index for title, parent_uid, and org_id" 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.289486854Z level=info msg="Migration successfully executed" id="Add unique index for title, parent_uid, and org_id" duration=600.644µs 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.290016685Z level=info msg="Executing migration" id="Sync dashboard and folder table" 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.290257928Z level=info msg="Migration successfully executed" id="Sync dashboard and folder table" duration=241.452µs 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.290685658Z level=info msg="Executing migration" id="Remove ghost folders from the folder table" 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.290815832Z level=info msg="Migration successfully executed" id="Remove ghost folders from the folder table" duration=130.413µs 2026-03-10T11:32:40.416 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.291288787Z level=info msg="Executing migration" id="Remove unique index UQE_folder_uid_org_id" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.291929145Z level=info msg="Migration successfully executed" id="Remove unique index UQE_folder_uid_org_id" duration=639.436µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.29248715Z level=info msg="Executing migration" id="Add unique index UQE_folder_org_id_uid" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.292959704Z level=info msg="Migration successfully executed" id="Add unique index UQE_folder_org_id_uid" duration=473.055µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.293392224Z level=info msg="Executing migration" id="Remove unique index UQE_folder_title_parent_uid_org_id" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.293837988Z level=info msg="Migration successfully executed" id="Remove unique index UQE_folder_title_parent_uid_org_id" duration=445.513µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.294319149Z level=info msg="Executing migration" id="Add unique index UQE_folder_org_id_parent_uid_title" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.294771876Z level=info msg="Migration successfully executed" id="Add unique index UQE_folder_org_id_parent_uid_title" duration=453.618µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.295247557Z level=info msg="Executing migration" id="Remove index IDX_folder_parent_uid_org_id" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.295957485Z level=info msg="Migration successfully executed" id="Remove index IDX_folder_parent_uid_org_id" duration=707.474µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.296526761Z level=info msg="Executing migration" id="create anon_device table" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.296974969Z level=info msg="Migration successfully executed" id="create anon_device table" duration=448.559µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.29742354Z level=info msg="Executing migration" id="add unique index anon_device.device_id" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.297950226Z level=info msg="Migration successfully executed" id="add unique index anon_device.device_id" duration=526.356µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.298437477Z level=info msg="Executing migration" id="add index anon_device.updated_at" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.298862573Z level=info msg="Migration successfully executed" id="add index anon_device.updated_at" duration=425.205µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.299303659Z level=info msg="Executing migration" id="create signing_key table" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.299730226Z level=info msg="Migration successfully executed" id="create signing_key table" duration=426.708µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.300540894Z level=info msg="Executing migration" id="add unique index signing_key.key_id" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.301154183Z level=info msg="Migration successfully executed" id="add unique index signing_key.key_id" duration=613.278µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.301729008Z level=info msg="Executing migration" id="set legacy alert migration status in kvstore" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.303804072Z level=info msg="Migration successfully executed" id="set legacy alert migration status in kvstore" duration=2.075085ms 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.304481169Z level=info msg="Executing migration" id="migrate record of created folders during legacy migration to kvstore" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.304720679Z level=info msg="Migration successfully executed" id="migrate record of created folders during legacy migration to kvstore" duration=239.87µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.305324498Z level=info msg="Executing migration" id="Add folder_uid for dashboard" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.308042355Z level=info msg="Migration successfully executed" id="Add folder_uid for dashboard" duration=2.717817ms 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.308651385Z level=info msg="Executing migration" id="Populate dashboard folder_uid column" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.309072112Z level=info msg="Migration successfully executed" id="Populate dashboard folder_uid column" duration=421.379µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.309644905Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_folder_uid_title" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.31017135Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_folder_uid_title" duration=526.314µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.310702735Z level=info msg="Executing migration" id="Delete unique index for dashboard_org_id_folder_id_title" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.311232566Z level=info msg="Migration successfully executed" id="Delete unique index for dashboard_org_id_folder_id_title" duration=529.581µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.311779651Z level=info msg="Executing migration" id="Delete unique index for dashboard_org_id_folder_uid_title" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.312289614Z level=info msg="Migration successfully executed" id="Delete unique index for dashboard_org_id_folder_uid_title" duration=509.704µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.312803247Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_folder_uid_title_is_folder" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.313347626Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_folder_uid_title_is_folder" duration=544.129µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.313890331Z level=info msg="Executing migration" id="Restore index for dashboard_org_id_folder_id_title" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.314416115Z level=info msg="Migration successfully executed" id="Restore index for dashboard_org_id_folder_id_title" duration=525.575µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.314932762Z level=info msg="Executing migration" id="create sso_setting table" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.315425515Z level=info msg="Migration successfully executed" id="create sso_setting table" duration=492.943µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.316023864Z level=info msg="Executing migration" id="copy kvstore migration status to each org" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.316499606Z level=info msg="Migration successfully executed" id="copy kvstore migration status to each org" duration=476.252µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.317060765Z level=info msg="Executing migration" id="add back entry for orgid=0 migrated status" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.317251152Z level=info msg="Migration successfully executed" id="add back entry for orgid=0 migrated status" duration=190.698µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.317851064Z level=info msg="Executing migration" id="alter kv_store.value to longtext" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.317947164Z level=info msg="Migration successfully executed" id="alter kv_store.value to longtext" duration=96.57µs 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.31848949Z level=info msg="Executing migration" id="add notification_settings column to alert_rule table" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.321221634Z level=info msg="Migration successfully executed" id="add notification_settings column to alert_rule table" duration=2.732145ms 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.321772054Z level=info msg="Executing migration" id="add notification_settings column to alert_rule_version table" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.324409019Z level=info msg="Migration successfully executed" id="add notification_settings column to alert_rule_version table" duration=2.636826ms 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.324964079Z level=info msg="Executing migration" id="removing scope from alert.instances:read action migration" 2026-03-10T11:32:40.417 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.325201322Z level=info msg="Migration successfully executed" id="removing scope from alert.instances:read action migration" duration=236.943µs 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=migrator t=2026-03-10T11:32:40.325734981Z level=info msg="migrations completed" performed=547 skipped=0 duration=763.797159ms 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=sqlstore t=2026-03-10T11:32:40.326406678Z level=info msg="Created default organization" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=secrets t=2026-03-10T11:32:40.327124001Z level=info msg="Envelope encryption state" enabled=true currentprovider=secretKey.v1 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=plugin.store t=2026-03-10T11:32:40.334952465Z level=info msg="Loading plugins..." 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=local.finder t=2026-03-10T11:32:40.372969593Z level=warn msg="Skipping finding plugins as directory does not exist" path=/usr/share/grafana/plugins-bundled 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=plugin.store t=2026-03-10T11:32:40.373166181Z level=info msg="Plugins loaded" count=55 duration=38.214068ms 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=query_data t=2026-03-10T11:32:40.374618309Z level=info msg="Query Service initialization" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=live.push_http t=2026-03-10T11:32:40.376183649Z level=info msg="Live Push Gateway initialization" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ngalert.migration t=2026-03-10T11:32:40.377487409Z level=info msg=Starting 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ngalert.migration t=2026-03-10T11:32:40.377765941Z level=info msg="Applying transition" currentType=Legacy desiredType=UnifiedAlerting cleanOnDowngrade=false cleanOnUpgrade=false 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ngalert.migration orgID=1 t=2026-03-10T11:32:40.378036047Z level=info msg="Migrating alerts for organisation" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ngalert.migration orgID=1 t=2026-03-10T11:32:40.378473886Z level=info msg="Alerts found to migrate" alerts=0 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ngalert.migration t=2026-03-10T11:32:40.379374882Z level=info msg="Completed alerting migration" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ngalert.state.manager t=2026-03-10T11:32:40.38640965Z level=info msg="Running in alternative execution of Error/NoData mode" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=infra.usagestats.collector t=2026-03-10T11:32:40.387500502Z level=info msg="registering usage stat providers" usageStatsProvidersLen=2 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=provisioning.datasources t=2026-03-10T11:32:40.388797509Z level=info msg="inserting datasource from configuration" name=Dashboard1 uid=P43CA22E17D0F9596 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=provisioning.datasources t=2026-03-10T11:32:40.393461308Z level=info msg="inserting datasource from configuration" name=Loki uid=P8E80F9AEF21F6940 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=provisioning.alerting t=2026-03-10T11:32:40.398322898Z level=info msg="starting to provision alerting" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=provisioning.alerting t=2026-03-10T11:32:40.398421813Z level=info msg="finished to provision alerting" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=http.server t=2026-03-10T11:32:40.399630746Z level=info msg="HTTP Server TLS settings" MinTLSVersion=TLS1.2 configuredciphers=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA,TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA,TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA,TLS_RSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_256_GCM_SHA384,TLS_RSA_WITH_AES_128_CBC_SHA,TLS_RSA_WITH_AES_256_CBC_SHA 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=http.server t=2026-03-10T11:32:40.399948852Z level=info msg="HTTP Server Listen" address=[::]:3000 protocol=https subUrl= socket= 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ngalert.state.manager t=2026-03-10T11:32:40.400031707Z level=info msg="Warming state cache for startup" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ngalert.state.manager t=2026-03-10T11:32:40.400243563Z level=info msg="State cache has been initialized" states=0 duration=212.086µs 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=provisioning.dashboard t=2026-03-10T11:32:40.404091024Z level=info msg="starting to provision dashboards" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=sqlstore.transactions t=2026-03-10T11:32:40.410538383Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=0 code="database is locked" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=grafanaStorageLogger t=2026-03-10T11:32:40.399959422Z level=info msg="Storage starting" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ngalert.multiorg.alertmanager t=2026-03-10T11:32:40.41196919Z level=info msg="Starting MultiOrg Alertmanager" 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ngalert.scheduler t=2026-03-10T11:32:40.412020176Z level=info msg="Starting scheduler" tickInterval=10s maxAttempts=1 2026-03-10T11:32:40.418 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ticker t=2026-03-10T11:32:40.412070751Z level=info msg=starting first_tick=2026-03-10T11:32:50Z 2026-03-10T11:32:40.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:40 vm06 ceph-mon[49228]: pgmap v10: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 0 B/s wr, 6 op/s 2026-03-10T11:32:40.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:40 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:40 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:40 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:40 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:40 vm06 ceph-mon[57405]: pgmap v10: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 0 B/s wr, 6 op/s 2026-03-10T11:32:40.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:40 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:40 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:40 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:40 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:40.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:40 vm06 bash[87146]: Trying to pull quay.io/prometheus/node-exporter:v1.7.0... 2026-03-10T11:32:40.730 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=sqlstore.transactions t=2026-03-10T11:32:40.417069567Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=0 code="database is locked" 2026-03-10T11:32:40.730 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=sqlstore.transactions t=2026-03-10T11:32:40.431962293Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=1 code="database is locked" 2026-03-10T11:32:40.730 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=sqlstore.transactions t=2026-03-10T11:32:40.432700084Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=1 code="database is locked" 2026-03-10T11:32:40.730 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=sqlstore.transactions t=2026-03-10T11:32:40.4427743Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=2 code="database is locked" 2026-03-10T11:32:40.730 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=sqlstore.transactions t=2026-03-10T11:32:40.443093358Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=2 code="database is locked" 2026-03-10T11:32:40.730 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=plugins.update.checker t=2026-03-10T11:32:40.478979167Z level=info msg="Update check succeeded" duration=67.899912ms 2026-03-10T11:32:40.730 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=provisioning.dashboard t=2026-03-10T11:32:40.52459101Z level=info msg="finished to provision dashboards" 2026-03-10T11:32:40.730 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=grafana-apiserver t=2026-03-10T11:32:40.58891044Z level=info msg="Adding GroupVersion playlist.grafana.app v0alpha1 to ResourceManager" 2026-03-10T11:32:40.730 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:32:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=grafana-apiserver t=2026-03-10T11:32:40.589503922Z level=info msg="Adding GroupVersion featuretoggle.grafana.app v0alpha1 to ResourceManager" 2026-03-10T11:32:41.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:40 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:32:40] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:32:41.428 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:41 vm06 ceph-mon[49228]: Deploying daemon node-exporter.a on vm06 2026-03-10T11:32:41.428 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:41 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:32:41.428 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:41 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:41.428 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:41 vm06 ceph-mon[57405]: Deploying daemon node-exporter.a on vm06 2026-03-10T11:32:41.428 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:41 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:32:41.428 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:41 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:41.428 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:41 vm06 bash[87146]: Getting image source signatures 2026-03-10T11:32:41.428 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:41 vm06 bash[87146]: Copying blob sha256:324153f2810a9927fcce320af9e4e291e0b6e805cbdd1f338386c756b9defa24 2026-03-10T11:32:41.428 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:41 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[86933]: ts=2026-03-10T11:32:41.306Z caller=cluster.go:698 level=info component=cluster msg="gossip settled; proceeding" elapsed=10.003965901s 2026-03-10T11:32:41.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:41 vm09 ceph-mon[54793]: Deploying daemon node-exporter.a on vm06 2026-03-10T11:32:41.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:41 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:32:41.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:41 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:41.781 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:41 vm06 bash[87146]: Copying blob sha256:2abcce694348cd2c949c0e98a7400ebdfd8341021bcf6b541bc72033ce982510 2026-03-10T11:32:41.781 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:41 vm06 bash[87146]: Copying blob sha256:455fd88e5221bc1e278ef2d059cd70e4df99a24e5af050ede621534276f6cf9a 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 bash[87146]: Copying config sha256:72c9c208898624938c9e4183d6686ea4a5fd3f912bc29bc3f00147924c521a3e 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 bash[87146]: Writing manifest to image destination 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 podman[87146]: 2026-03-10 11:32:42.274624444 +0000 UTC m=+2.067259344 container create 2e000946ad6d135b7409d27a9ce0a2256b01388f61b9ef719a26be6c508cdcf9 (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 podman[87146]: 2026-03-10 11:32:42.314023894 +0000 UTC m=+2.106658794 container init 2e000946ad6d135b7409d27a9ce0a2256b01388f61b9ef719a26be6c508cdcf9 (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 podman[87146]: 2026-03-10 11:32:42.31932701 +0000 UTC m=+2.111961910 container start 2e000946ad6d135b7409d27a9ce0a2256b01388f61b9ef719a26be6c508cdcf9 (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 bash[87146]: 2e000946ad6d135b7409d27a9ce0a2256b01388f61b9ef719a26be6c508cdcf9 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 podman[87146]: 2026-03-10 11:32:42.265633755 +0000 UTC m=+2.058268664 image pull 72c9c208898624938c9e4183d6686ea4a5fd3f912bc29bc3f00147924c521a3e quay.io/prometheus/node-exporter:v1.7.0 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.322Z caller=node_exporter.go:192 level=info msg="Starting node_exporter" version="(version=1.7.0, branch=HEAD, revision=7333465abf9efba81876303bb57e6fadb946041b)" 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.323Z caller=node_exporter.go:193 level=info msg="Build context" build_context="(go=go1.21.4, platform=linux/amd64, user=root@35918982f6d8, date=20231112-23:53:35, tags=netgo osusergo static_build)" 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.323Z caller=diskstats_common.go:111 level=info collector=diskstats msg="Parsed flag --collector.diskstats.device-exclude" flag=^(ram|loop|fd|(h|s|v|xv)d[a-z]|nvme\d+n\d+p)\d+$ 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.323Z caller=diskstats_linux.go:265 level=error collector=diskstats msg="Failed to open directory, disabling udev device properties" path=/run/udev/data 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=filesystem_common.go:111 level=info collector=filesystem msg="Parsed flag --collector.filesystem.mount-points-exclude" flag=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/containers/storage/.+)($|/) 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=filesystem_common.go:113 level=info collector=filesystem msg="Parsed flag --collector.filesystem.fs-types-exclude" flag=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:110 level=info msg="Enabled collectors" 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=arp 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=bcache 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=bonding 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=btrfs 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=conntrack 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=cpu 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=cpufreq 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=diskstats 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=dmi 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=edac 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=entropy 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=fibrechannel 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=filefd 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=filesystem 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=hwmon 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=infiniband 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=ipvs 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=loadavg 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=mdadm 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.324Z caller=node_exporter.go:117 level=info collector=meminfo 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=netclass 2026-03-10T11:32:42.532 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=netdev 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=netstat 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=nfs 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=nfsd 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=nvme 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=os 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=powersupplyclass 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=pressure 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=rapl 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=schedstat 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=selinux 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.325Z caller=node_exporter.go:117 level=info collector=sockstat 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 systemd[1]: Started Ceph node-exporter.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.327Z caller=node_exporter.go:117 level=info collector=softnet 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.327Z caller=node_exporter.go:117 level=info collector=stat 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.327Z caller=node_exporter.go:117 level=info collector=tapestats 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.327Z caller=node_exporter.go:117 level=info collector=textfile 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.327Z caller=node_exporter.go:117 level=info collector=thermal_zone 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.327Z caller=node_exporter.go:117 level=info collector=time 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.328Z caller=node_exporter.go:117 level=info collector=udp_queues 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.328Z caller=node_exporter.go:117 level=info collector=uname 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.328Z caller=node_exporter.go:117 level=info collector=vmstat 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.328Z caller=node_exporter.go:117 level=info collector=xfs 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.329Z caller=node_exporter.go:117 level=info collector=zfs 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.329Z caller=tls_config.go:274 level=info msg="Listening on" address=[::]:9100 2026-03-10T11:32:42.533 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a[87199]: ts=2026-03-10T11:32:42.329Z caller=tls_config.go:277 level=info msg="TLS is disabled." http2=false address=[::]:9100 2026-03-10T11:32:42.833 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:42 vm09 ceph-mon[54793]: pgmap v11: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 0 B/s wr, 6 op/s 2026-03-10T11:32:42.833 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:42 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:42.833 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:42 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:42.833 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:42 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:43.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-mon[49228]: pgmap v11: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 0 B/s wr, 6 op/s 2026-03-10T11:32:43.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:43.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:43.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:42 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:43.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:42 vm06 ceph-mon[57405]: pgmap v11: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 0 B/s wr, 6 op/s 2026-03-10T11:32:43.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:42 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:43.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:42 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:43.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:42 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:43.230 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:43 vm09 bash[81223]: Trying to pull quay.io/prometheus/node-exporter:v1.7.0... 2026-03-10T11:32:43.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:43 vm09 ceph-mon[54793]: Deploying daemon node-exporter.b on vm09 2026-03-10T11:32:43.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:43 vm09 ceph-mon[54793]: pgmap v12: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 16 KiB/s rd, 0 B/s wr, 7 op/s 2026-03-10T11:32:44.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:43 vm06 ceph-mon[49228]: Deploying daemon node-exporter.b on vm09 2026-03-10T11:32:44.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:43 vm06 ceph-mon[49228]: pgmap v12: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 16 KiB/s rd, 0 B/s wr, 7 op/s 2026-03-10T11:32:44.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:43 vm06 ceph-mon[57405]: Deploying daemon node-exporter.b on vm09 2026-03-10T11:32:44.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:43 vm06 ceph-mon[57405]: pgmap v12: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 16 KiB/s rd, 0 B/s wr, 7 op/s 2026-03-10T11:32:44.730 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:44 vm09 bash[81223]: Getting image source signatures 2026-03-10T11:32:44.730 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:44 vm09 bash[81223]: Copying blob sha256:324153f2810a9927fcce320af9e4e291e0b6e805cbdd1f338386c756b9defa24 2026-03-10T11:32:44.730 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:44 vm09 bash[81223]: Copying blob sha256:2abcce694348cd2c949c0e98a7400ebdfd8341021bcf6b541bc72033ce982510 2026-03-10T11:32:44.730 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:44 vm09 bash[81223]: Copying blob sha256:455fd88e5221bc1e278ef2d059cd70e4df99a24e5af050ede621534276f6cf9a 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:44 vm09 bash[81223]: Copying config sha256:72c9c208898624938c9e4183d6686ea4a5fd3f912bc29bc3f00147924c521a3e 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:44 vm09 bash[81223]: Writing manifest to image destination 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:44 vm09 podman[81223]: 2026-03-10 11:32:44.996551565 +0000 UTC m=+1.983367524 container create 3b0b72eff720abfb5b984cc56d49c55c2819e6b5d04e50b28ac6360ca1725dfd (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b, maintainer=The Prometheus Authors ) 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 podman[81223]: 2026-03-10 11:32:44.990433474 +0000 UTC m=+1.977249452 image pull 72c9c208898624938c9e4183d6686ea4a5fd3f912bc29bc3f00147924c521a3e quay.io/prometheus/node-exporter:v1.7.0 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 podman[81223]: 2026-03-10 11:32:45.037478711 +0000 UTC m=+2.024294680 container init 3b0b72eff720abfb5b984cc56d49c55c2819e6b5d04e50b28ac6360ca1725dfd (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b, maintainer=The Prometheus Authors ) 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 podman[81223]: 2026-03-10 11:32:45.040474347 +0000 UTC m=+2.027290327 container start 3b0b72eff720abfb5b984cc56d49c55c2819e6b5d04e50b28ac6360ca1725dfd (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b, maintainer=The Prometheus Authors ) 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 bash[81223]: 3b0b72eff720abfb5b984cc56d49c55c2819e6b5d04e50b28ac6360ca1725dfd 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 systemd[1]: Started Ceph node-exporter.b for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.047Z caller=node_exporter.go:192 level=info msg="Starting node_exporter" version="(version=1.7.0, branch=HEAD, revision=7333465abf9efba81876303bb57e6fadb946041b)" 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:193 level=info msg="Build context" build_context="(go=go1.21.4, platform=linux/amd64, user=root@35918982f6d8, date=20231112-23:53:35, tags=netgo osusergo static_build)" 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=diskstats_common.go:111 level=info collector=diskstats msg="Parsed flag --collector.diskstats.device-exclude" flag=^(ram|loop|fd|(h|s|v|xv)d[a-z]|nvme\d+n\d+p)\d+$ 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=diskstats_linux.go:265 level=error collector=diskstats msg="Failed to open directory, disabling udev device properties" path=/run/udev/data 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=filesystem_common.go:111 level=info collector=filesystem msg="Parsed flag --collector.filesystem.mount-points-exclude" flag=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/containers/storage/.+)($|/) 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=filesystem_common.go:113 level=info collector=filesystem msg="Parsed flag --collector.filesystem.fs-types-exclude" flag=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:110 level=info msg="Enabled collectors" 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=arp 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=bcache 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=bonding 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=btrfs 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=conntrack 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=cpu 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=cpufreq 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=diskstats 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=dmi 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=edac 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=entropy 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=fibrechannel 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=filefd 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=filesystem 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=hwmon 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=infiniband 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=ipvs 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=loadavg 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=mdadm 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=meminfo 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=netclass 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=netdev 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=netstat 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=nfs 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=nfsd 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=nvme 2026-03-10T11:32:45.446 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=os 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=powersupplyclass 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=pressure 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=rapl 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=schedstat 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=selinux 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=sockstat 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=softnet 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=stat 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=tapestats 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=textfile 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=thermal_zone 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=time 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=udp_queues 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=uname 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=vmstat 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=xfs 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.048Z caller=node_exporter.go:117 level=info collector=zfs 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.049Z caller=tls_config.go:274 level=info msg="Listening on" address=[::]:9100 2026-03-10T11:32:45.447 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:32:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b[81279]: ts=2026-03-10T11:32:45.049Z caller=tls_config.go:277 level=info msg="TLS is disabled." http2=false address=[::]:9100 2026-03-10T11:32:46.083 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:46 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:32:46.083 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:46 vm09 ceph-mon[54793]: pgmap v13: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:46.083 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:46 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.083 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:46 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.083 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:46 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.083 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:46 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.083 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:46 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.083 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:46 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:46.083 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:46 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[49228]: pgmap v13: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[57405]: pgmap v13: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:46.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:46.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:46 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.418 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:47.418 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.418 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.418 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.418 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.418 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:47.418 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:47.418 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.418 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 systemd[1]: Stopping Ceph alertmanager.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:32:47.419 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:47.419 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.419 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.419 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.419 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.419 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:47.419 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:47.419 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.670 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[86933]: ts=2026-03-10T11:32:47.417Z caller=main.go:583 level=info msg="Received SIGTERM, exiting gracefully..." 2026-03-10T11:32:47.670 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 podman[87698]: 2026-03-10 11:32:47.431627964 +0000 UTC m=+0.030299116 container died a392601a12a81aa54a722e8b98624ab9788ef048e4e9125c07c51a7c41c8a6e4 (image=quay.io/prometheus/alertmanager:v0.25.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:47.670 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 podman[87698]: 2026-03-10 11:32:47.454717084 +0000 UTC m=+0.053388236 container remove a392601a12a81aa54a722e8b98624ab9788ef048e4e9125c07c51a7c41c8a6e4 (image=quay.io/prometheus/alertmanager:v0.25.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:47.670 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 podman[87698]: 2026-03-10 11:32:47.45590598 +0000 UTC m=+0.054577132 volume remove 276c777a39590941ccba6155b4cd8b907a457841e4a1fedb30db4959fb50e3b0 2026-03-10T11:32:47.670 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 bash[87698]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a 2026-03-10T11:32:47.670 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@alertmanager.a.service: Deactivated successfully. 2026-03-10T11:32:47.670 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 systemd[1]: Stopped Ceph alertmanager.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:32:47.670 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 systemd[1]: Starting Ceph alertmanager.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:32:47.670 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 podman[87765]: 2026-03-10 11:32:47.642428584 +0000 UTC m=+0.017819818 volume create 9941b3ef7aacf5bbfcf28fe6b4c2b11321afebb37d561514847bbf20a14e4bbb 2026-03-10T11:32:47.670 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 podman[87765]: 2026-03-10 11:32:47.644792448 +0000 UTC m=+0.020183671 container create d067b405a2f9af608fc12e8da0d517c85068cd4673120d32c246b1e119cd414d (image=quay.io/prometheus/alertmanager:v0.25.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:47.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:47 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:47.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:47 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:47 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:47 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:47 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:47.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:47 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:47.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:47 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:47.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:47 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:48.031 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 podman[87765]: 2026-03-10 11:32:47.673895866 +0000 UTC m=+0.049287109 container init d067b405a2f9af608fc12e8da0d517c85068cd4673120d32c246b1e119cd414d (image=quay.io/prometheus/alertmanager:v0.25.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 podman[87765]: 2026-03-10 11:32:47.677576696 +0000 UTC m=+0.052967930 container start d067b405a2f9af608fc12e8da0d517c85068cd4673120d32c246b1e119cd414d (image=quay.io/prometheus/alertmanager:v0.25.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 bash[87765]: d067b405a2f9af608fc12e8da0d517c85068cd4673120d32c246b1e119cd414d 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 podman[87765]: 2026-03-10 11:32:47.635317412 +0000 UTC m=+0.010708646 image pull c8568f914cd25b2062c44e9f79f9c18da6e3b85fe0c47a12a2191c61426c2b19 quay.io/prometheus/alertmanager:v0.25.0 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 systemd[1]: Started Ceph alertmanager.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[87776]: ts=2026-03-10T11:32:47.696Z caller=main.go:240 level=info msg="Starting Alertmanager" version="(version=0.25.0, branch=HEAD, revision=258fab7cdd551f2cf251ed0348f0ad7289aee789)" 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[87776]: ts=2026-03-10T11:32:47.696Z caller=main.go:241 level=info build_context="(go=go1.19.4, user=root@abe866dd5717, date=20221222-14:51:36)" 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[87776]: ts=2026-03-10T11:32:47.697Z caller=cluster.go:185 level=info component=cluster msg="setting advertise address explicitly" addr=192.168.123.106 port=9094 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[87776]: ts=2026-03-10T11:32:47.700Z caller=cluster.go:681 level=info component=cluster msg="Waiting for gossip to settle..." interval=2s 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[87776]: ts=2026-03-10T11:32:47.739Z caller=coordinator.go:113 level=info component=configuration msg="Loading configuration file" file=/etc/alertmanager/alertmanager.yml 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[87776]: ts=2026-03-10T11:32:47.740Z caller=coordinator.go:126 level=info component=configuration msg="Completed loading of configuration file" file=/etc/alertmanager/alertmanager.yml 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[87776]: ts=2026-03-10T11:32:47.746Z caller=tls_config.go:232 level=info msg="Listening on" address=[::]:9093 2026-03-10T11:32:48.032 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[87776]: ts=2026-03-10T11:32:47.747Z caller=tls_config.go:235 level=info msg="TLS is disabled." http2=false address=[::]:9093 2026-03-10T11:32:48.342 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:48 vm09 ceph-mon[54793]: Reconfiguring alertmanager.a (dependencies changed)... 2026-03-10T11:32:48.342 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:48 vm09 ceph-mon[54793]: Reconfiguring daemon alertmanager.a on vm06 2026-03-10T11:32:48.342 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:48 vm09 ceph-mon[54793]: pgmap v14: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:32:48.342 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:48 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:48.342 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:48 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:48.342 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 systemd[1]: Stopping Ceph prometheus.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:32:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:48 vm06 ceph-mon[49228]: Reconfiguring alertmanager.a (dependencies changed)... 2026-03-10T11:32:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:48 vm06 ceph-mon[49228]: Reconfiguring daemon alertmanager.a on vm06 2026-03-10T11:32:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:48 vm06 ceph-mon[49228]: pgmap v14: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:32:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:48 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:48 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:48 vm06 ceph-mon[57405]: Reconfiguring alertmanager.a (dependencies changed)... 2026-03-10T11:32:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:48 vm06 ceph-mon[57405]: Reconfiguring daemon alertmanager.a on vm06 2026-03-10T11:32:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:48 vm06 ceph-mon[57405]: pgmap v14: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:32:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:48 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:48 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:48.596 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.342Z caller=main.go:964 level=warn msg="Received SIGTERM, exiting gracefully..." 2026-03-10T11:32:48.596 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.342Z caller=main.go:988 level=info msg="Stopping scrape discovery manager..." 2026-03-10T11:32:48.596 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.342Z caller=main.go:1002 level=info msg="Stopping notify discovery manager..." 2026-03-10T11:32:48.596 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.342Z caller=manager.go:177 level=info component="rule manager" msg="Stopping rule manager..." 2026-03-10T11:32:48.596 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.344Z caller=manager.go:187 level=info component="rule manager" msg="Rule manager stopped" 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.344Z caller=main.go:1039 level=info msg="Stopping scrape manager..." 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.344Z caller=main.go:984 level=info msg="Scrape discovery manager stopped" 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.344Z caller=main.go:998 level=info msg="Notify discovery manager stopped" 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.344Z caller=main.go:1031 level=info msg="Scrape manager stopped" 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.345Z caller=notifier.go:618 level=info component=notifier msg="Stopping notification manager..." 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.345Z caller=main.go:1261 level=info msg="Notifier manager stopped" 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[79497]: ts=2026-03-10T11:32:48.345Z caller=main.go:1273 level=info msg="See you next time!" 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 podman[81832]: 2026-03-10 11:32:48.356852759 +0000 UTC m=+0.029609824 container died b914580585251e79c1fb0ea6ce1c4175355c80a6d1df2396e806728636ee5251 (image=quay.io/prometheus/prometheus:v2.51.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 podman[81832]: 2026-03-10 11:32:48.382670974 +0000 UTC m=+0.055428039 container remove b914580585251e79c1fb0ea6ce1c4175355c80a6d1df2396e806728636ee5251 (image=quay.io/prometheus/prometheus:v2.51.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 bash[81832]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@prometheus.a.service: Deactivated successfully. 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 systemd[1]: Stopped Ceph prometheus.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 systemd[1]: Starting Ceph prometheus.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:32:48.597 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 podman[81902]: 2026-03-10 11:32:48.570795716 +0000 UTC m=+0.025592744 container create 817d17d9e41df65713d81e7cb9cacde905bea7df323f4ef983c7a058fd8de4e5 (image=quay.io/prometheus/prometheus:v2.51.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:48.948 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:48] ENGINE Bus STOPPING 2026-03-10T11:32:48.948 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:48] ENGINE HTTP Server cherrypy._cpwsgi_server.CPWSGIServer(('::', 9283)) shut down 2026-03-10T11:32:48.948 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:48] ENGINE Bus STOPPED 2026-03-10T11:32:48.948 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:48] ENGINE Bus STARTING 2026-03-10T11:32:48.949 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:48] ENGINE Serving on http://:::9283 2026-03-10T11:32:48.949 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:48] ENGINE Bus STARTED 2026-03-10T11:32:48.949 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:48] ENGINE Bus STOPPING 2026-03-10T11:32:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 podman[81902]: 2026-03-10 11:32:48.600751036 +0000 UTC m=+0.055548074 container init 817d17d9e41df65713d81e7cb9cacde905bea7df323f4ef983c7a058fd8de4e5 (image=quay.io/prometheus/prometheus:v2.51.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 podman[81902]: 2026-03-10 11:32:48.603979007 +0000 UTC m=+0.058776035 container start 817d17d9e41df65713d81e7cb9cacde905bea7df323f4ef983c7a058fd8de4e5 (image=quay.io/prometheus/prometheus:v2.51.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a, maintainer=The Prometheus Authors ) 2026-03-10T11:32:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 bash[81902]: 817d17d9e41df65713d81e7cb9cacde905bea7df323f4ef983c7a058fd8de4e5 2026-03-10T11:32:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 podman[81902]: 2026-03-10 11:32:48.562482817 +0000 UTC m=+0.017279845 image pull 1d3b7f56885b6dd623f1785be963aa9c195f86bc256ea454e8d02a7980b79c53 quay.io/prometheus/prometheus:v2.51.0 2026-03-10T11:32:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 systemd[1]: Started Ceph prometheus.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:32:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.635Z caller=main.go:617 level=info msg="Starting Prometheus Server" mode=server version="(version=2.51.0, branch=HEAD, revision=c05c15512acb675e3f6cd662a6727854e93fc024)" 2026-03-10T11:32:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.635Z caller=main.go:622 level=info build_context="(go=go1.22.1, platform=linux/amd64, user=root@b5723e458358, date=20240319-10:54:45, tags=netgo,builtinassets,stringlabels)" 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.635Z caller=main.go:623 level=info host_details="(Linux 5.14.0-686.el9.x86_64 #1 SMP PREEMPT_DYNAMIC Thu Feb 19 10:49:27 UTC 2026 x86_64 vm09 (none))" 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.635Z caller=main.go:624 level=info fd_limits="(soft=1048576, hard=1048576)" 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.635Z caller=main.go:625 level=info vm_limits="(soft=unlimited, hard=unlimited)" 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.640Z caller=web.go:568 level=info component=web msg="Start listening for connections" address=:9095 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.640Z caller=main.go:1129 level=info msg="Starting TSDB ..." 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.642Z caller=head.go:616 level=info component=tsdb msg="Replaying on-disk memory mappable chunks if any" 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.642Z caller=head.go:698 level=info component=tsdb msg="On-disk memory mappable chunks replay completed" duration=1.172µs 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.642Z caller=head.go:706 level=info component=tsdb msg="Replaying WAL, this may take a while" 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.642Z caller=tls_config.go:313 level=info component=web msg="Listening on" address=[::]:9095 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.642Z caller=tls_config.go:316 level=info component=web msg="TLS is disabled." http2=false address=[::]:9095 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.643Z caller=head.go:778 level=info component=tsdb msg="WAL segment loaded" segment=0 maxSegment=1 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.645Z caller=head.go:778 level=info component=tsdb msg="WAL segment loaded" segment=1 maxSegment=1 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.645Z caller=head.go:815 level=info component=tsdb msg="WAL replay completed" checkpoint_replay_duration=155.772µs wal_replay_duration=3.150286ms wbl_replay_duration=129ns total_replay_duration=3.31807ms 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.646Z caller=main.go:1150 level=info fs_type=XFS_SUPER_MAGIC 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.646Z caller=main.go:1153 level=info msg="TSDB started" 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.647Z caller=main.go:1335 level=info msg="Loading configuration file" filename=/etc/prometheus/prometheus.yml 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.657Z caller=main.go:1372 level=info msg="Completed loading of configuration file" filename=/etc/prometheus/prometheus.yml totalDuration=10.643141ms db_storage=921ns remote_storage=1.443µs web_handler=381ns query_engine=852ns scrape=724.756µs scrape_sd=86.884µs notify=8.004µs notify_sd=22.562µs rules=8.39324ms tracing=6.001µs 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.657Z caller=main.go:1114 level=info msg="Server is ready to receive web requests." 2026-03-10T11:32:48.981 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:32:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:32:48.657Z caller=manager.go:163 level=info component="rule manager" msg="Starting rule manager..." 2026-03-10T11:32:49.205 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:48] ENGINE HTTP Server cherrypy._cpwsgi_server.CPWSGIServer(('::', 9283)) shut down 2026-03-10T11:32:49.206 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:48] ENGINE Bus STOPPED 2026-03-10T11:32:49.206 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:48] ENGINE Bus STARTING 2026-03-10T11:32:49.206 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:49] ENGINE Serving on http://:::9283 2026-03-10T11:32:49.206 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:49] ENGINE Bus STARTED 2026-03-10T11:32:49.206 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:49] ENGINE Bus STOPPING 2026-03-10T11:32:49.408 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: Reconfiguring prometheus.a (dependencies changed)... 2026-03-10T11:32:49.408 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: Reconfiguring daemon prometheus.a on vm09 2026-03-10T11:32:49.408 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.409 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.409 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T11:32:49.409 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm06.local:9093"}]: dispatch 2026-03-10T11:32:49.409 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.409 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T11:32:49.409 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm09.local:9095"}]: dispatch 2026-03-10T11:32:49.409 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.409 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T11:32:49.409 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm09.local:3000"}]: dispatch 2026-03-10T11:32:49.409 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.409 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:49 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: Reconfiguring prometheus.a (dependencies changed)... 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: Reconfiguring daemon prometheus.a on vm09 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm06.local:9093"}]: dispatch 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm09.local:9095"}]: dispatch 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm09.local:3000"}]: dispatch 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: Reconfiguring prometheus.a (dependencies changed)... 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: Reconfiguring daemon prometheus.a on vm09 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm06.local:9093"}]: dispatch 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm09.local:9095"}]: dispatch 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm09.local:3000"}]: dispatch 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:49.532 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:49 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:32:49.822 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[87776]: ts=2026-03-10T11:32:49.700Z caller=cluster.go:706 level=info component=cluster msg="gossip not settled" polls=0 before=0 now=1 elapsed=2.000829044s 2026-03-10T11:32:49.822 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:49] ENGINE HTTP Server cherrypy._cpwsgi_server.CPWSGIServer(('::', 9283)) shut down 2026-03-10T11:32:49.822 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:49] ENGINE Bus STOPPED 2026-03-10T11:32:49.822 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:49] ENGINE Bus STARTING 2026-03-10T11:32:49.822 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:49] ENGINE Serving on http://:::9283 2026-03-10T11:32:49.822 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:32:49 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:32:49] ENGINE Bus STARTED 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm06.local:9093"}]: dispatch 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm09.local:9095"}]: dispatch 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm09.local:3000"}]: dispatch 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: pgmap v15: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:50 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T11:32:50.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm06.local:9093"}]: dispatch 2026-03-10T11:32:50.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T11:32:50.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm09.local:9095"}]: dispatch 2026-03-10T11:32:50.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T11:32:50.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm09.local:3000"}]: dispatch 2026-03-10T11:32:50.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: pgmap v15: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:50.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm06.local:9093"}]: dispatch 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm09.local:9095"}]: dispatch 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mon.? -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm09.local:3000"}]: dispatch 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: pgmap v15: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:32:50.782 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:50 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:32:52.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:52 vm09 ceph-mon[54793]: pgmap v16: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:52 vm06 ceph-mon[49228]: pgmap v16: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:52.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:52 vm06 ceph-mon[57405]: pgmap v16: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:54.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:54 vm09 ceph-mon[54793]: pgmap v17: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:32:54.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:54 vm06 ceph-mon[49228]: pgmap v17: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:32:54.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:54 vm06 ceph-mon[57405]: pgmap v17: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:32:55.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:55 vm09 ceph-mon[54793]: pgmap v18: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:55.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:55 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:32:56.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:55 vm06 ceph-mon[49228]: pgmap v18: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:56.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:55 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:32:56.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:55 vm06 ceph-mon[57405]: pgmap v18: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:32:56.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:55 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:32:56.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:32:56 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:32:56.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:56 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:57.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:56 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:57.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:56 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:32:57.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:32:57 vm09 ceph-mon[54793]: pgmap v19: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:32:58.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:32:57 vm06 ceph-mon[49228]: pgmap v19: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:32:58.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:32:57 vm06 ceph-mon[57405]: pgmap v19: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:32:58.031 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:32:57 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[87776]: ts=2026-03-10T11:32:57.704Z caller=cluster.go:698 level=info component=cluster msg="gossip settled; proceeding" elapsed=10.003935532s 2026-03-10T11:33:00.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:00 vm09 ceph-mon[54793]: pgmap v20: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:00.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:00 vm06 ceph-mon[49228]: pgmap v20: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:00.563 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:00 vm06 ceph-mon[57405]: pgmap v20: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:01.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:33:00 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:33:00] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:33:02.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:02 vm09 ceph-mon[54793]: pgmap v21: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:02.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:02 vm06 ceph-mon[49228]: pgmap v21: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:02.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:02 vm06 ceph-mon[57405]: pgmap v21: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:04.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:04 vm09 ceph-mon[54793]: pgmap v22: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:04.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:04 vm06 ceph-mon[49228]: pgmap v22: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:04.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:04 vm06 ceph-mon[57405]: pgmap v22: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:06.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:06 vm09 ceph-mon[54793]: pgmap v23: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:06.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:33:06 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:33:06.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:06 vm06 ceph-mon[49228]: pgmap v23: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:06.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:06 vm06 ceph-mon[57405]: pgmap v23: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:07.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:07 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:07.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:07 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:07.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:07 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:08.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:08 vm09 ceph-mon[54793]: pgmap v24: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:08.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:08 vm06 ceph-mon[49228]: pgmap v24: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:08.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:08 vm06 ceph-mon[57405]: pgmap v24: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:10.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:10 vm09 ceph-mon[54793]: pgmap v25: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:10.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:10 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:10.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:10 vm06 ceph-mon[49228]: pgmap v25: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:10.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:10 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:10.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:10 vm06 ceph-mon[57405]: pgmap v25: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:10.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:10 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:11.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:33:10 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:33:10] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:33:12.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:12 vm06 ceph-mon[49228]: pgmap v26: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:12.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:12 vm06 ceph-mon[57405]: pgmap v26: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:12.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:12 vm09 ceph-mon[54793]: pgmap v26: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:14.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:14 vm06 ceph-mon[49228]: pgmap v27: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:14.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:14 vm06 ceph-mon[57405]: pgmap v27: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:14.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:14 vm09 ceph-mon[54793]: pgmap v27: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:16.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:16 vm09 ceph-mon[54793]: pgmap v28: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:16.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:33:16 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:33:16.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:16 vm06 ceph-mon[49228]: pgmap v28: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:16.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:16 vm06 ceph-mon[57405]: pgmap v28: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:17.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:17 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:17.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:17 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:17.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:17 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:18.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:18 vm09 ceph-mon[54793]: pgmap v29: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:18.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:18 vm06 ceph-mon[49228]: pgmap v29: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:18.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:18 vm06 ceph-mon[57405]: pgmap v29: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:19.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:19 vm09 ceph-mon[54793]: pgmap v30: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:19.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:19 vm06 ceph-mon[49228]: pgmap v30: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:19.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:19 vm06 ceph-mon[57405]: pgmap v30: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:20.730 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:33:20 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=infra.usagestats t=2026-03-10T11:33:20.416647905Z level=info msg="Usage stats are ready to report" 2026-03-10T11:33:21.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:33:20 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:33:20] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:33:21.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:21 vm09 ceph-mon[54793]: pgmap v31: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:22.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:21 vm06 ceph-mon[49228]: pgmap v31: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:22.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:21 vm06 ceph-mon[57405]: pgmap v31: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:24.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:24 vm09 ceph-mon[54793]: pgmap v32: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:24.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:24 vm06 ceph-mon[49228]: pgmap v32: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:24.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:24 vm06 ceph-mon[57405]: pgmap v32: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:25.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.13", "id": [1, 5]}]: dispatch 2026-03-10T11:33:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.13", "id": [1, 5]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.13", "id": [1, 5]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.13", "id": [1, 5]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.13", "id": [1, 5]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]: dispatch 2026-03-10T11:33:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.13", "id": [1, 5]}]: dispatch 2026-03-10T11:33:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:26 vm09 ceph-mon[54793]: pgmap v33: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:26 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]': finished 2026-03-10T11:33:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:26 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]': finished 2026-03-10T11:33:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:26 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.13", "id": [1, 5]}]': finished 2026-03-10T11:33:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:26 vm09 ceph-mon[54793]: osdmap e58: 8 total, 8 up, 8 in 2026-03-10T11:33:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:26 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:26 vm09 ceph-mon[54793]: osdmap e59: 8 total, 8 up, 8 in 2026-03-10T11:33:26.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:33:26 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[49228]: pgmap v33: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]': finished 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]': finished 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.13", "id": [1, 5]}]': finished 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[49228]: osdmap e58: 8 total, 8 up, 8 in 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[49228]: osdmap e59: 8 total, 8 up, 8 in 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[57405]: pgmap v33: 132 pgs: 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]': finished 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.7", "id": [1, 2]}]': finished 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.13", "id": [1, 5]}]': finished 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[57405]: osdmap e58: 8 total, 8 up, 8 in 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:26 vm06 ceph-mon[57405]: osdmap e59: 8 total, 8 up, 8 in 2026-03-10T11:33:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:27 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:27.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:27 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:27.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:27 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:28 vm09 ceph-mon[54793]: pgmap v36: 132 pgs: 3 peering, 129 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1 op/s 2026-03-10T11:33:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:28 vm09 ceph-mon[54793]: Health check failed: Reduced data availability: 3 pgs peering (PG_AVAILABILITY) 2026-03-10T11:33:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:28 vm06 ceph-mon[49228]: pgmap v36: 132 pgs: 3 peering, 129 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1 op/s 2026-03-10T11:33:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:28 vm06 ceph-mon[49228]: Health check failed: Reduced data availability: 3 pgs peering (PG_AVAILABILITY) 2026-03-10T11:33:28.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:28 vm06 ceph-mon[57405]: pgmap v36: 132 pgs: 3 peering, 129 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1 op/s 2026-03-10T11:33:28.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:28 vm06 ceph-mon[57405]: Health check failed: Reduced data availability: 3 pgs peering (PG_AVAILABILITY) 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr:Note: switching to '75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b'. 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr: 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr:state without impacting any branches by switching back to a branch. 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr: 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr: 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr: git switch -c 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr: 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr:Or undo this operation with: 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr: 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr: git switch - 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr: 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr: 2026-03-10T11:33:29.556 INFO:tasks.workunit.client.0.vm06.stderr:HEAD is now at 75a68fd8ca3 qa/suites/orch/cephadm/osds: drop nvme_loop task 2026-03-10T11:33:29.562 DEBUG:teuthology.orchestra.run.vm06:> cd -- /home/ubuntu/cephtest/clone.client.0/qa/workunits && if test -e Makefile ; then make ; fi && find -executable -type f -printf '%P\0' >/home/ubuntu/cephtest/workunits.list.client.0 2026-03-10T11:33:29.622 INFO:tasks.workunit.client.0.vm06.stdout:for d in direct_io fs ; do ( cd $d ; make all ) ; done 2026-03-10T11:33:29.624 INFO:tasks.workunit.client.0.vm06.stdout:make[1]: Entering directory '/home/ubuntu/cephtest/clone.client.0/qa/workunits/direct_io' 2026-03-10T11:33:29.624 INFO:tasks.workunit.client.0.vm06.stdout:cc -Wall -Wextra -D_GNU_SOURCE direct_io_test.c -o direct_io_test 2026-03-10T11:33:29.666 INFO:tasks.workunit.client.0.vm06.stdout:cc -Wall -Wextra -D_GNU_SOURCE test_sync_io.c -o test_sync_io 2026-03-10T11:33:29.699 INFO:tasks.workunit.client.0.vm06.stdout:cc -Wall -Wextra -D_GNU_SOURCE test_short_dio_read.c -o test_short_dio_read 2026-03-10T11:33:29.730 INFO:tasks.workunit.client.0.vm06.stdout:make[1]: Leaving directory '/home/ubuntu/cephtest/clone.client.0/qa/workunits/direct_io' 2026-03-10T11:33:29.731 INFO:tasks.workunit.client.0.vm06.stdout:make[1]: Entering directory '/home/ubuntu/cephtest/clone.client.0/qa/workunits/fs' 2026-03-10T11:33:29.731 INFO:tasks.workunit.client.0.vm06.stdout:cc -Wall -Wextra -D_GNU_SOURCE test_o_trunc.c -o test_o_trunc 2026-03-10T11:33:29.761 INFO:tasks.workunit.client.0.vm06.stdout:make[1]: Leaving directory '/home/ubuntu/cephtest/clone.client.0/qa/workunits/fs' 2026-03-10T11:33:29.764 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T11:33:29.764 DEBUG:teuthology.orchestra.run.vm06:> dd if=/home/ubuntu/cephtest/workunits.list.client.0 of=/dev/stdout 2026-03-10T11:33:29.822 INFO:tasks.workunit:Running workunits matching rados/test_python.sh on client.0... 2026-03-10T11:33:29.823 INFO:tasks.workunit:Running workunit rados/test_python.sh... 2026-03-10T11:33:29.823 DEBUG:teuthology.orchestra.run.vm06:workunit test rados/test_python.sh> mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 1h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test_python.sh 2026-03-10T11:33:29.885 INFO:tasks.workunit.client.0.vm06.stderr:+ ceph osd pool create rbd 2026-03-10T11:33:30.247 INFO:tasks.workunit.client.0.vm06.stderr:pool 'rbd' already exists 2026-03-10T11:33:30.257 INFO:tasks.workunit.client.0.vm06.stderr:++ dirname /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test_python.sh 2026-03-10T11:33:30.258 INFO:tasks.workunit.client.0.vm06.stderr:+ python3 -m pytest -v /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/../../../src/test/pybind/test_rados.py 2026-03-10T11:33:30.345 INFO:tasks.workunit.client.0.vm06.stdout:============================= test session starts ============================== 2026-03-10T11:33:30.345 INFO:tasks.workunit.client.0.vm06.stdout:platform linux -- Python 3.9.25, pytest-6.2.2, py-1.10.0, pluggy-0.13.1 -- /usr/bin/python3 2026-03-10T11:33:30.345 INFO:tasks.workunit.client.0.vm06.stdout:cachedir: .pytest_cache 2026-03-10T11:33:30.345 INFO:tasks.workunit.client.0.vm06.stdout:rootdir: /home/ubuntu/cephtest/clone.client.0/src/test/pybind, configfile: pytest.ini 2026-03-10T11:33:30.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:30 vm09 ceph-mon[54793]: pgmap v37: 132 pgs: 3 peering, 129 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1 op/s 2026-03-10T11:33:30.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:30 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1900050221' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "rbd"}]: dispatch 2026-03-10T11:33:30.500 INFO:tasks.workunit.client.0.vm06.stdout:collecting ... collected 91 items 2026-03-10T11:33:30.501 INFO:tasks.workunit.client.0.vm06.stdout: 2026-03-10T11:33:30.506 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::test_rados_init_error PASSED [ 1%] 2026-03-10T11:33:30.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:30 vm06 ceph-mon[49228]: pgmap v37: 132 pgs: 3 peering, 129 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1 op/s 2026-03-10T11:33:30.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:30 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1900050221' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "rbd"}]: dispatch 2026-03-10T11:33:30.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:30 vm06 ceph-mon[57405]: pgmap v37: 132 pgs: 3 peering, 129 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1 op/s 2026-03-10T11:33:30.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:30 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1900050221' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "rbd"}]: dispatch 2026-03-10T11:33:30.543 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::test_rados_init PASSED [ 2%] 2026-03-10T11:33:30.553 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::test_ioctx_context_manager PASSED [ 3%] 2026-03-10T11:33:30.560 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::test_parse_argv PASSED [ 4%] 2026-03-10T11:33:30.563 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::test_parse_argv_empty_str PASSED [ 5%] 2026-03-10T11:33:30.568 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRadosStateError::test_configuring PASSED [ 6%] 2026-03-10T11:33:30.578 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRadosStateError::test_connected PASSED [ 7%] 2026-03-10T11:33:30.587 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRadosStateError::test_shutdown PASSED [ 8%] 2026-03-10T11:33:30.602 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_ping_monitor PASSED [ 9%] 2026-03-10T11:33:30.615 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_annotations PASSED [ 10%] 2026-03-10T11:33:31.030 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:33:30 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:33:30] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:33:31.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:31 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1900050221' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "rbd"}]': finished 2026-03-10T11:33:31.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:31 vm06 ceph-mon[49228]: osdmap e60: 8 total, 8 up, 8 in 2026-03-10T11:33:31.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:31 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1900050221' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "rbd"}]: dispatch 2026-03-10T11:33:31.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:31 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3508888224' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:33:31.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:31 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1900050221' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "rbd"}]': finished 2026-03-10T11:33:31.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:31 vm06 ceph-mon[57405]: osdmap e60: 8 total, 8 up, 8 in 2026-03-10T11:33:31.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:31 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1900050221' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "rbd"}]: dispatch 2026-03-10T11:33:31.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:31 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3508888224' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:33:31.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:31 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1900050221' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "rbd"}]': finished 2026-03-10T11:33:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:31 vm09 ceph-mon[54793]: osdmap e60: 8 total, 8 up, 8 in 2026-03-10T11:33:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:31 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1900050221' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "rbd"}]: dispatch 2026-03-10T11:33:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:31 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3508888224' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:33:32.218 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_create PASSED [ 12%] 2026-03-10T11:33:32.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:32 vm06 ceph-mon[49228]: pgmap v39: 164 pgs: 7 creating+peering, 25 unknown, 3 peering, 129 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 681 B/s rd, 0 op/s 2026-03-10T11:33:32.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:32 vm06 ceph-mon[49228]: osdmap e61: 8 total, 8 up, 8 in 2026-03-10T11:33:32.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:32 vm06 ceph-mon[49228]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:32.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:32 vm06 ceph-mon[49228]: osdmap e62: 8 total, 8 up, 8 in 2026-03-10T11:33:32.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:32 vm06 ceph-mon[57405]: pgmap v39: 164 pgs: 7 creating+peering, 25 unknown, 3 peering, 129 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 681 B/s rd, 0 op/s 2026-03-10T11:33:32.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:32 vm06 ceph-mon[57405]: osdmap e61: 8 total, 8 up, 8 in 2026-03-10T11:33:32.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:32 vm06 ceph-mon[57405]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:32.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:32 vm06 ceph-mon[57405]: osdmap e62: 8 total, 8 up, 8 in 2026-03-10T11:33:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:32 vm09 ceph-mon[54793]: pgmap v39: 164 pgs: 7 creating+peering, 25 unknown, 3 peering, 129 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 681 B/s rd, 0 op/s 2026-03-10T11:33:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:32 vm09 ceph-mon[54793]: osdmap e61: 8 total, 8 up, 8 in 2026-03-10T11:33:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:32 vm09 ceph-mon[54793]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:32 vm09 ceph-mon[54793]: osdmap e62: 8 total, 8 up, 8 in 2026-03-10T11:33:34.219 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_create_utf8 PASSED [ 13%] 2026-03-10T11:33:34.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:34 vm06 ceph-mon[49228]: pgmap v42: 164 pgs: 32 creating+peering, 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s; 73 B/s, 2 objects/s recovering 2026-03-10T11:33:34.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:34 vm06 ceph-mon[49228]: osdmap e63: 8 total, 8 up, 8 in 2026-03-10T11:33:34.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:34 vm06 ceph-mon[49228]: Health check cleared: PG_AVAILABILITY (was: Reduced data availability: 3 pgs peering) 2026-03-10T11:33:34.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:34 vm06 ceph-mon[57405]: pgmap v42: 164 pgs: 32 creating+peering, 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s; 73 B/s, 2 objects/s recovering 2026-03-10T11:33:34.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:34 vm06 ceph-mon[57405]: osdmap e63: 8 total, 8 up, 8 in 2026-03-10T11:33:34.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:34 vm06 ceph-mon[57405]: Health check cleared: PG_AVAILABILITY (was: Reduced data availability: 3 pgs peering) 2026-03-10T11:33:34.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:34 vm09 ceph-mon[54793]: pgmap v42: 164 pgs: 32 creating+peering, 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 853 B/s rd, 0 op/s; 73 B/s, 2 objects/s recovering 2026-03-10T11:33:34.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:34 vm09 ceph-mon[54793]: osdmap e63: 8 total, 8 up, 8 in 2026-03-10T11:33:34.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:34 vm09 ceph-mon[54793]: Health check cleared: PG_AVAILABILITY (was: Reduced data availability: 3 pgs peering) 2026-03-10T11:33:35.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:35 vm06 ceph-mon[49228]: osdmap e64: 8 total, 8 up, 8 in 2026-03-10T11:33:35.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:35 vm06 ceph-mon[57405]: osdmap e64: 8 total, 8 up, 8 in 2026-03-10T11:33:35.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:35 vm09 ceph-mon[54793]: osdmap e64: 8 total, 8 up, 8 in 2026-03-10T11:33:36.268 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_pool_lookup_utf8 PASSED [ 14%] 2026-03-10T11:33:36.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:36 vm09 ceph-mon[54793]: pgmap v45: 164 pgs: 32 creating+peering, 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s; 109 B/s, 3 objects/s recovering 2026-03-10T11:33:36.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:36 vm09 ceph-mon[54793]: osdmap e65: 8 total, 8 up, 8 in 2026-03-10T11:33:36.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:33:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:33:36.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:36 vm06 ceph-mon[49228]: pgmap v45: 164 pgs: 32 creating+peering, 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s; 109 B/s, 3 objects/s recovering 2026-03-10T11:33:36.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:36 vm06 ceph-mon[49228]: osdmap e65: 8 total, 8 up, 8 in 2026-03-10T11:33:36.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:36 vm06 ceph-mon[57405]: pgmap v45: 164 pgs: 32 creating+peering, 132 active+clean; 455 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s; 109 B/s, 3 objects/s recovering 2026-03-10T11:33:36.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:36 vm06 ceph-mon[57405]: osdmap e65: 8 total, 8 up, 8 in 2026-03-10T11:33:37.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:37 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:37.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:37 vm06 ceph-mon[49228]: osdmap e66: 8 total, 8 up, 8 in 2026-03-10T11:33:37.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:37 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:37.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:37 vm06 ceph-mon[57405]: osdmap e66: 8 total, 8 up, 8 in 2026-03-10T11:33:37.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:37 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:37.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:37 vm09 ceph-mon[54793]: osdmap e66: 8 total, 8 up, 8 in 2026-03-10T11:33:38.298 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_eexist PASSED [ 15%] 2026-03-10T11:33:38.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:38 vm06 ceph-mon[49228]: pgmap v48: 164 pgs: 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:38.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:38 vm06 ceph-mon[49228]: osdmap e67: 8 total, 8 up, 8 in 2026-03-10T11:33:38.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:38 vm06 ceph-mon[57405]: pgmap v48: 164 pgs: 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:38.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:38 vm06 ceph-mon[57405]: osdmap e67: 8 total, 8 up, 8 in 2026-03-10T11:33:38.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:38 vm09 ceph-mon[54793]: pgmap v48: 164 pgs: 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:38.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:38 vm09 ceph-mon[54793]: osdmap e67: 8 total, 8 up, 8 in 2026-03-10T11:33:39.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:39 vm09 ceph-mon[54793]: osdmap e68: 8 total, 8 up, 8 in 2026-03-10T11:33:39.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:39 vm06 ceph-mon[49228]: osdmap e68: 8 total, 8 up, 8 in 2026-03-10T11:33:39.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:39 vm06 ceph-mon[57405]: osdmap e68: 8 total, 8 up, 8 in 2026-03-10T11:33:40.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:40 vm09 ceph-mon[54793]: pgmap v51: 164 pgs: 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:40.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:40 vm09 ceph-mon[54793]: osdmap e69: 8 total, 8 up, 8 in 2026-03-10T11:33:40.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:40 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:40.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:40 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:40.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:40 vm06 ceph-mon[49228]: pgmap v51: 164 pgs: 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:40.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:40 vm06 ceph-mon[49228]: osdmap e69: 8 total, 8 up, 8 in 2026-03-10T11:33:40.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:40 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:40.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:40 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:40.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:40 vm06 ceph-mon[57405]: pgmap v51: 164 pgs: 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:40.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:40 vm06 ceph-mon[57405]: osdmap e69: 8 total, 8 up, 8 in 2026-03-10T11:33:40.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:40 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:40.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:40 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:41.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:33:40 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:33:40] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:33:41.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:41 vm09 ceph-mon[54793]: osdmap e70: 8 total, 8 up, 8 in 2026-03-10T11:33:41.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:41 vm06 ceph-mon[49228]: osdmap e70: 8 total, 8 up, 8 in 2026-03-10T11:33:41.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:41 vm06 ceph-mon[57405]: osdmap e70: 8 total, 8 up, 8 in 2026-03-10T11:33:42.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:42 vm09 ceph-mon[54793]: pgmap v54: 228 pgs: 3 creating+peering, 54 unknown, 171 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:33:42.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:42 vm09 ceph-mon[54793]: Health check update: 3 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:42.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:42 vm09 ceph-mon[54793]: osdmap e71: 8 total, 8 up, 8 in 2026-03-10T11:33:42.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:42 vm09 ceph-mon[54793]: osdmap e72: 8 total, 8 up, 8 in 2026-03-10T11:33:42.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:42 vm06 ceph-mon[49228]: pgmap v54: 228 pgs: 3 creating+peering, 54 unknown, 171 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:33:42.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:42 vm06 ceph-mon[49228]: Health check update: 3 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:42.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:42 vm06 ceph-mon[49228]: osdmap e71: 8 total, 8 up, 8 in 2026-03-10T11:33:42.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:42 vm06 ceph-mon[49228]: osdmap e72: 8 total, 8 up, 8 in 2026-03-10T11:33:42.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:42 vm06 ceph-mon[57405]: pgmap v54: 228 pgs: 3 creating+peering, 54 unknown, 171 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:33:42.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:42 vm06 ceph-mon[57405]: Health check update: 3 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:42.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:42 vm06 ceph-mon[57405]: osdmap e71: 8 total, 8 up, 8 in 2026-03-10T11:33:42.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:42 vm06 ceph-mon[57405]: osdmap e72: 8 total, 8 up, 8 in 2026-03-10T11:33:43.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:43 vm09 ceph-mon[54793]: pgmap v57: 228 pgs: 32 unknown, 32 creating+peering, 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:43.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:43 vm09 ceph-mon[54793]: osdmap e73: 8 total, 8 up, 8 in 2026-03-10T11:33:43.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:43 vm06 ceph-mon[49228]: pgmap v57: 228 pgs: 32 unknown, 32 creating+peering, 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:43.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:43 vm06 ceph-mon[49228]: osdmap e73: 8 total, 8 up, 8 in 2026-03-10T11:33:43.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:43 vm06 ceph-mon[57405]: pgmap v57: 228 pgs: 32 unknown, 32 creating+peering, 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:43.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:43 vm06 ceph-mon[57405]: osdmap e73: 8 total, 8 up, 8 in 2026-03-10T11:33:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:45 vm09 ceph-mon[54793]: osdmap e74: 8 total, 8 up, 8 in 2026-03-10T11:33:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:45 vm09 ceph-mon[54793]: pgmap v60: 164 pgs: 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:45.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:45 vm06 ceph-mon[49228]: osdmap e74: 8 total, 8 up, 8 in 2026-03-10T11:33:45.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:45 vm06 ceph-mon[49228]: pgmap v60: 164 pgs: 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:45.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:45 vm06 ceph-mon[57405]: osdmap e74: 8 total, 8 up, 8 in 2026-03-10T11:33:45.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:45 vm06 ceph-mon[57405]: pgmap v60: 164 pgs: 164 active+clean; 455 KiB data, 217 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:46.412 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_list_pools PASSED [ 16%] 2026-03-10T11:33:46.425 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:33:46 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:33:46.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:46 vm09 ceph-mon[54793]: osdmap e75: 8 total, 8 up, 8 in 2026-03-10T11:33:46.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:46 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:46.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:46 vm06 ceph-mon[49228]: osdmap e75: 8 total, 8 up, 8 in 2026-03-10T11:33:46.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:46 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:46.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:46 vm06 ceph-mon[57405]: osdmap e75: 8 total, 8 up, 8 in 2026-03-10T11:33:46.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:46 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:47.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:47 vm06 ceph-mon[49228]: osdmap e76: 8 total, 8 up, 8 in 2026-03-10T11:33:47.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:47 vm06 ceph-mon[49228]: pgmap v63: 164 pgs: 164 active+clean; 455 KiB data, 222 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:47.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:47 vm06 ceph-mon[57405]: osdmap e76: 8 total, 8 up, 8 in 2026-03-10T11:33:47.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:47 vm06 ceph-mon[57405]: pgmap v63: 164 pgs: 164 active+clean; 455 KiB data, 222 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:47.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:47 vm09 ceph-mon[54793]: osdmap e76: 8 total, 8 up, 8 in 2026-03-10T11:33:47.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:47 vm09 ceph-mon[54793]: pgmap v63: 164 pgs: 164 active+clean; 455 KiB data, 222 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:48.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:48 vm09 ceph-mon[54793]: osdmap e77: 8 total, 8 up, 8 in 2026-03-10T11:33:48.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:48 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:49.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:48 vm06 ceph-mon[49228]: osdmap e77: 8 total, 8 up, 8 in 2026-03-10T11:33:49.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:48 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:49.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:48 vm06 ceph-mon[57405]: osdmap e77: 8 total, 8 up, 8 in 2026-03-10T11:33:49.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:48 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:49.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:49 vm09 ceph-mon[54793]: osdmap e78: 8 total, 8 up, 8 in 2026-03-10T11:33:49.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:49 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1391328641' entity='client.admin' cmd=[{"prefix": "osd tier add", "pool": "foo", "tierpool": "foo-cache", "force_nonempty": ""}]: dispatch 2026-03-10T11:33:49.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:49 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd tier add", "pool": "foo", "tierpool": "foo-cache", "force_nonempty": ""}]: dispatch 2026-03-10T11:33:49.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:49 vm09 ceph-mon[54793]: pgmap v66: 228 pgs: 64 unknown, 164 active+clean; 455 KiB data, 222 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:49.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:49 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd tier add", "pool": "foo", "tierpool": "foo-cache", "force_nonempty": ""}]': finished 2026-03-10T11:33:49.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:49 vm09 ceph-mon[54793]: osdmap e79: 8 total, 8 up, 8 in 2026-03-10T11:33:49.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:49 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1391328641' entity='client.admin' cmd=[{"prefix": "osd tier cache-mode", "pool": "foo-cache", "tierpool": "foo-cache", "mode": "readonly", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:33:49.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:49 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd tier cache-mode", "pool": "foo-cache", "tierpool": "foo-cache", "mode": "readonly", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[49228]: osdmap e78: 8 total, 8 up, 8 in 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1391328641' entity='client.admin' cmd=[{"prefix": "osd tier add", "pool": "foo", "tierpool": "foo-cache", "force_nonempty": ""}]: dispatch 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd tier add", "pool": "foo", "tierpool": "foo-cache", "force_nonempty": ""}]: dispatch 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[49228]: pgmap v66: 228 pgs: 64 unknown, 164 active+clean; 455 KiB data, 222 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd tier add", "pool": "foo", "tierpool": "foo-cache", "force_nonempty": ""}]': finished 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[49228]: osdmap e79: 8 total, 8 up, 8 in 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1391328641' entity='client.admin' cmd=[{"prefix": "osd tier cache-mode", "pool": "foo-cache", "tierpool": "foo-cache", "mode": "readonly", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd tier cache-mode", "pool": "foo-cache", "tierpool": "foo-cache", "mode": "readonly", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[57405]: osdmap e78: 8 total, 8 up, 8 in 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1391328641' entity='client.admin' cmd=[{"prefix": "osd tier add", "pool": "foo", "tierpool": "foo-cache", "force_nonempty": ""}]: dispatch 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd tier add", "pool": "foo", "tierpool": "foo-cache", "force_nonempty": ""}]: dispatch 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[57405]: pgmap v66: 228 pgs: 64 unknown, 164 active+clean; 455 KiB data, 222 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd tier add", "pool": "foo", "tierpool": "foo-cache", "force_nonempty": ""}]': finished 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[57405]: osdmap e79: 8 total, 8 up, 8 in 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1391328641' entity='client.admin' cmd=[{"prefix": "osd tier cache-mode", "pool": "foo-cache", "tierpool": "foo-cache", "mode": "readonly", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:33:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:49 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd tier cache-mode", "pool": "foo-cache", "tierpool": "foo-cache", "mode": "readonly", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:33:50.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:50 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:33:50.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:50 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd tier cache-mode", "pool": "foo-cache", "tierpool": "foo-cache", "mode": "readonly", "yes_i_really_mean_it": true}]': finished 2026-03-10T11:33:50.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:50 vm09 ceph-mon[54793]: osdmap e80: 8 total, 8 up, 8 in 2026-03-10T11:33:50.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:50 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1391328641' entity='client.admin' cmd=[{"prefix": "osd tier remove", "pool": "foo", "tierpool": "foo-cache"}]: dispatch 2026-03-10T11:33:50.731 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:50 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd tier remove", "pool": "foo", "tierpool": "foo-cache"}]: dispatch 2026-03-10T11:33:50.744 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:33:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:33:50] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:33:50.744 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:50 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:33:50.744 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:50 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd tier cache-mode", "pool": "foo-cache", "tierpool": "foo-cache", "mode": "readonly", "yes_i_really_mean_it": true}]': finished 2026-03-10T11:33:50.744 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:50 vm06 ceph-mon[49228]: osdmap e80: 8 total, 8 up, 8 in 2026-03-10T11:33:50.745 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:50 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1391328641' entity='client.admin' cmd=[{"prefix": "osd tier remove", "pool": "foo", "tierpool": "foo-cache"}]: dispatch 2026-03-10T11:33:50.745 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:50 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd tier remove", "pool": "foo", "tierpool": "foo-cache"}]: dispatch 2026-03-10T11:33:50.745 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:50 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:33:50.745 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:50 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd tier cache-mode", "pool": "foo-cache", "tierpool": "foo-cache", "mode": "readonly", "yes_i_really_mean_it": true}]': finished 2026-03-10T11:33:50.745 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:50 vm06 ceph-mon[57405]: osdmap e80: 8 total, 8 up, 8 in 2026-03-10T11:33:50.745 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:50 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1391328641' entity='client.admin' cmd=[{"prefix": "osd tier remove", "pool": "foo", "tierpool": "foo-cache"}]: dispatch 2026-03-10T11:33:50.745 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:50 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd tier remove", "pool": "foo", "tierpool": "foo-cache"}]: dispatch 2026-03-10T11:33:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:51 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:51 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:51 vm09 ceph-mon[54793]: pgmap v69: 228 pgs: 51 unknown, 177 active+clean; 455 KiB data, 222 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:33:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:51 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:51 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:51 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:33:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:51 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:33:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:51 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:51 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd tier remove", "pool": "foo", "tierpool": "foo-cache"}]': finished 2026-03-10T11:33:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:51 vm09 ceph-mon[54793]: osdmap e81: 8 total, 8 up, 8 in 2026-03-10T11:33:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[49228]: pgmap v69: 228 pgs: 51 unknown, 177 active+clean; 455 KiB data, 222 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:33:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:33:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:33:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd tier remove", "pool": "foo", "tierpool": "foo-cache"}]': finished 2026-03-10T11:33:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[49228]: osdmap e81: 8 total, 8 up, 8 in 2026-03-10T11:33:52.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[57405]: pgmap v69: 228 pgs: 51 unknown, 177 active+clean; 455 KiB data, 222 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:33:52.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:33:52.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:33:52.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:33:52.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd tier remove", "pool": "foo", "tierpool": "foo-cache"}]': finished 2026-03-10T11:33:52.282 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:51 vm06 ceph-mon[57405]: osdmap e81: 8 total, 8 up, 8 in 2026-03-10T11:33:53.787 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_get_pool_base_tier PASSED [ 17%] 2026-03-10T11:33:53.801 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_get_fsid PASSED [ 18%] 2026-03-10T11:33:54.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:53 vm06 ceph-mon[49228]: osdmap e82: 8 total, 8 up, 8 in 2026-03-10T11:33:54.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:53 vm06 ceph-mon[49228]: pgmap v72: 196 pgs: 196 active+clean; 455 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:54.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:53 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:54.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:53 vm06 ceph-mon[57405]: osdmap e82: 8 total, 8 up, 8 in 2026-03-10T11:33:54.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:53 vm06 ceph-mon[57405]: pgmap v72: 196 pgs: 196 active+clean; 455 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:54.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:53 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:54.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:53 vm09 ceph-mon[54793]: osdmap e82: 8 total, 8 up, 8 in 2026-03-10T11:33:54.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:53 vm09 ceph-mon[54793]: pgmap v72: 196 pgs: 196 active+clean; 455 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:54.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:53 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:33:54.796 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_blocklist_add PASSED [ 19%] 2026-03-10T11:33:54.807 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_get_cluster_stats PASSED [ 20%] 2026-03-10T11:33:54.819 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestRados::test_monitor_log PASSED [ 21%] 2026-03-10T11:33:55.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:54 vm09 ceph-mon[54793]: osdmap e83: 8 total, 8 up, 8 in 2026-03-10T11:33:55.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:54 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1697861517' entity='client.admin' cmd=[{"prefix": "osd blocklist", "blocklistop": "add", "addr": "1.2.3.4/123", "expire": 1.0}]: dispatch 2026-03-10T11:33:55.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:54 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd blocklist", "blocklistop": "add", "addr": "1.2.3.4/123", "expire": 1.0}]: dispatch 2026-03-10T11:33:55.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:54 vm06 ceph-mon[49228]: osdmap e83: 8 total, 8 up, 8 in 2026-03-10T11:33:55.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:54 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1697861517' entity='client.admin' cmd=[{"prefix": "osd blocklist", "blocklistop": "add", "addr": "1.2.3.4/123", "expire": 1.0}]: dispatch 2026-03-10T11:33:55.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:54 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd blocklist", "blocklistop": "add", "addr": "1.2.3.4/123", "expire": 1.0}]: dispatch 2026-03-10T11:33:55.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:54 vm06 ceph-mon[57405]: osdmap e83: 8 total, 8 up, 8 in 2026-03-10T11:33:55.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:54 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1697861517' entity='client.admin' cmd=[{"prefix": "osd blocklist", "blocklistop": "add", "addr": "1.2.3.4/123", "expire": 1.0}]: dispatch 2026-03-10T11:33:55.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:54 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd blocklist", "blocklistop": "add", "addr": "1.2.3.4/123", "expire": 1.0}]: dispatch 2026-03-10T11:33:56.125 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:55 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd blocklist", "blocklistop": "add", "addr": "1.2.3.4/123", "expire": 1.0}]': finished 2026-03-10T11:33:56.125 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:55 vm09 ceph-mon[54793]: osdmap e84: 8 total, 8 up, 8 in 2026-03-10T11:33:56.125 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:55 vm09 ceph-mon[54793]: pgmap v75: 164 pgs: 164 active+clean; 455 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:56.125 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:55 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:56.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:55 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd blocklist", "blocklistop": "add", "addr": "1.2.3.4/123", "expire": 1.0}]': finished 2026-03-10T11:33:56.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:55 vm06 ceph-mon[49228]: osdmap e84: 8 total, 8 up, 8 in 2026-03-10T11:33:56.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:55 vm06 ceph-mon[49228]: pgmap v75: 164 pgs: 164 active+clean; 455 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:56.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:55 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:56.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:55 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd blocklist", "blocklistop": "add", "addr": "1.2.3.4/123", "expire": 1.0}]': finished 2026-03-10T11:33:56.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:55 vm06 ceph-mon[57405]: osdmap e84: 8 total, 8 up, 8 in 2026-03-10T11:33:56.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:55 vm06 ceph-mon[57405]: pgmap v75: 164 pgs: 164 active+clean; 455 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:56.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:55 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:33:56.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:33:56 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:33:57.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:56 vm09 ceph-mon[54793]: osdmap e85: 8 total, 8 up, 8 in 2026-03-10T11:33:57.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:56 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3625159867' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:33:57.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:56 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:33:57.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:56 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:57.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:56 vm06 ceph-mon[49228]: osdmap e85: 8 total, 8 up, 8 in 2026-03-10T11:33:57.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:56 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3625159867' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:33:57.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:56 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:33:57.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:56 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:57.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:56 vm06 ceph-mon[57405]: osdmap e85: 8 total, 8 up, 8 in 2026-03-10T11:33:57.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:56 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3625159867' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:33:57.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:56 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:33:57.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:56 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:33:57.824 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_get_last_version PASSED [ 23%] 2026-03-10T11:33:58.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:57 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:33:58.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:57 vm09 ceph-mon[54793]: osdmap e86: 8 total, 8 up, 8 in 2026-03-10T11:33:58.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:57 vm09 ceph-mon[54793]: pgmap v78: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 227 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:58.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:57 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:33:58.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:57 vm06 ceph-mon[49228]: osdmap e86: 8 total, 8 up, 8 in 2026-03-10T11:33:58.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:57 vm06 ceph-mon[49228]: pgmap v78: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 227 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:58.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:57 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:33:58.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:57 vm06 ceph-mon[57405]: osdmap e86: 8 total, 8 up, 8 in 2026-03-10T11:33:58.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:57 vm06 ceph-mon[57405]: pgmap v78: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 227 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:59.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:58 vm09 ceph-mon[54793]: osdmap e87: 8 total, 8 up, 8 in 2026-03-10T11:33:59.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:58 vm09 ceph-mon[54793]: osdmap e88: 8 total, 8 up, 8 in 2026-03-10T11:33:59.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:58 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/280656083' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:33:59.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:58 vm06 ceph-mon[49228]: osdmap e87: 8 total, 8 up, 8 in 2026-03-10T11:33:59.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:58 vm06 ceph-mon[49228]: osdmap e88: 8 total, 8 up, 8 in 2026-03-10T11:33:59.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:58 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/280656083' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:33:59.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:58 vm06 ceph-mon[57405]: osdmap e87: 8 total, 8 up, 8 in 2026-03-10T11:33:59.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:58 vm06 ceph-mon[57405]: osdmap e88: 8 total, 8 up, 8 in 2026-03-10T11:33:59.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:58 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/280656083' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:33:59.826 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:59 vm06 ceph-mon[49228]: pgmap v81: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 227 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:33:59.826 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:59 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/280656083' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:33:59.826 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:33:59 vm06 ceph-mon[49228]: osdmap e89: 8 total, 8 up, 8 in 2026-03-10T11:33:59.826 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:59 vm06 ceph-mon[57405]: pgmap v81: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 227 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:00.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:59 vm09 ceph-mon[54793]: pgmap v81: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 227 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:00.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:59 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/280656083' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:00.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:33:59 vm09 ceph-mon[54793]: osdmap e89: 8 total, 8 up, 8 in 2026-03-10T11:34:00.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:59 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/280656083' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:00.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:33:59 vm06 ceph-mon[57405]: osdmap e89: 8 total, 8 up, 8 in 2026-03-10T11:34:00.453 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_get_stats PASSED [ 24%] 2026-03-10T11:34:01.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:34:00 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:34:00] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:34:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:01 vm09 ceph-mon[54793]: osdmap e90: 8 total, 8 up, 8 in 2026-03-10T11:34:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:01 vm09 ceph-mon[54793]: pgmap v84: 164 pgs: 164 active+clean; 455 KiB data, 227 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:01.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:01 vm06 ceph-mon[49228]: osdmap e90: 8 total, 8 up, 8 in 2026-03-10T11:34:01.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:01 vm06 ceph-mon[49228]: pgmap v84: 164 pgs: 164 active+clean; 455 KiB data, 227 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:01.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:01 vm06 ceph-mon[57405]: osdmap e90: 8 total, 8 up, 8 in 2026-03-10T11:34:01.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:01 vm06 ceph-mon[57405]: pgmap v84: 164 pgs: 164 active+clean; 455 KiB data, 227 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:02.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:02 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:02.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:02 vm06 ceph-mon[49228]: osdmap e91: 8 total, 8 up, 8 in 2026-03-10T11:34:02.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:02 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:02.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:02 vm06 ceph-mon[57405]: osdmap e91: 8 total, 8 up, 8 in 2026-03-10T11:34:02.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:02 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:02.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:02 vm09 ceph-mon[54793]: osdmap e91: 8 total, 8 up, 8 in 2026-03-10T11:34:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:03 vm09 ceph-mon[54793]: osdmap e92: 8 total, 8 up, 8 in 2026-03-10T11:34:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:03 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1557440113' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:03 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:03 vm09 ceph-mon[54793]: pgmap v87: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 258 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:04.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:03 vm06 ceph-mon[49228]: osdmap e92: 8 total, 8 up, 8 in 2026-03-10T11:34:04.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:03 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1557440113' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:04.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:03 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:04.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:03 vm06 ceph-mon[49228]: pgmap v87: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 258 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:04.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:03 vm06 ceph-mon[57405]: osdmap e92: 8 total, 8 up, 8 in 2026-03-10T11:34:04.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:03 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1557440113' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:04.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:03 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:04.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:03 vm06 ceph-mon[57405]: pgmap v87: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 258 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:04.535 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_write PASSED [ 25%] 2026-03-10T11:34:04.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:04 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:04.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:04 vm09 ceph-mon[54793]: osdmap e93: 8 total, 8 up, 8 in 2026-03-10T11:34:04.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:04 vm09 ceph-mon[54793]: osdmap e94: 8 total, 8 up, 8 in 2026-03-10T11:34:05.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:04 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:05.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:04 vm06 ceph-mon[49228]: osdmap e93: 8 total, 8 up, 8 in 2026-03-10T11:34:05.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:04 vm06 ceph-mon[49228]: osdmap e94: 8 total, 8 up, 8 in 2026-03-10T11:34:05.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:04 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:05.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:04 vm06 ceph-mon[57405]: osdmap e93: 8 total, 8 up, 8 in 2026-03-10T11:34:05.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:04 vm06 ceph-mon[57405]: osdmap e94: 8 total, 8 up, 8 in 2026-03-10T11:34:05.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:05 vm09 ceph-mon[54793]: pgmap v90: 164 pgs: 164 active+clean; 455 KiB data, 258 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:06.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:05 vm06 ceph-mon[49228]: pgmap v90: 164 pgs: 164 active+clean; 455 KiB data, 258 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:06.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:05 vm06 ceph-mon[57405]: pgmap v90: 164 pgs: 164 active+clean; 455 KiB data, 258 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:06.479 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:34:06 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:34:06.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:06 vm09 ceph-mon[54793]: osdmap e95: 8 total, 8 up, 8 in 2026-03-10T11:34:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:06 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:07.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:06 vm06 ceph-mon[49228]: osdmap e95: 8 total, 8 up, 8 in 2026-03-10T11:34:07.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:06 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:07.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:06 vm06 ceph-mon[57405]: osdmap e95: 8 total, 8 up, 8 in 2026-03-10T11:34:07.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:06 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:07 vm09 ceph-mon[54793]: osdmap e96: 8 total, 8 up, 8 in 2026-03-10T11:34:08.036 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:07 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3604014233' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:08.036 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:07 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:08.036 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:07 vm09 ceph-mon[54793]: pgmap v93: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 276 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:08.036 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:07 vm06 ceph-mon[49228]: osdmap e96: 8 total, 8 up, 8 in 2026-03-10T11:34:08.036 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:07 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3604014233' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:08.036 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:07 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:08.037 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:07 vm06 ceph-mon[49228]: pgmap v93: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 276 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:08.037 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:07 vm06 ceph-mon[57405]: osdmap e96: 8 total, 8 up, 8 in 2026-03-10T11:34:08.037 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:07 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3604014233' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:08.037 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:07 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:08.037 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:07 vm06 ceph-mon[57405]: pgmap v93: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 276 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:08.679 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_write_full PASSED [ 26%] 2026-03-10T11:34:08.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:08 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:08 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:08 vm09 ceph-mon[54793]: osdmap e97: 8 total, 8 up, 8 in 2026-03-10T11:34:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:08 vm09 ceph-mon[54793]: osdmap e98: 8 total, 8 up, 8 in 2026-03-10T11:34:09.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:08 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:09.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:08 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:09.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:08 vm06 ceph-mon[49228]: osdmap e97: 8 total, 8 up, 8 in 2026-03-10T11:34:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:08 vm06 ceph-mon[49228]: osdmap e98: 8 total, 8 up, 8 in 2026-03-10T11:34:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:08 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:08 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:08 vm06 ceph-mon[57405]: osdmap e97: 8 total, 8 up, 8 in 2026-03-10T11:34:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:08 vm06 ceph-mon[57405]: osdmap e98: 8 total, 8 up, 8 in 2026-03-10T11:34:09.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:09 vm09 ceph-mon[54793]: pgmap v96: 164 pgs: 164 active+clean; 455 KiB data, 276 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:10.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:09 vm06 ceph-mon[49228]: pgmap v96: 164 pgs: 164 active+clean; 455 KiB data, 276 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:09 vm06 ceph-mon[57405]: pgmap v96: 164 pgs: 164 active+clean; 455 KiB data, 276 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:10.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:10 vm09 ceph-mon[54793]: osdmap e99: 8 total, 8 up, 8 in 2026-03-10T11:34:10.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:10 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:11.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:10 vm06 ceph-mon[49228]: osdmap e99: 8 total, 8 up, 8 in 2026-03-10T11:34:11.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:10 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:11.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:34:10 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:34:10] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:34:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:10 vm06 ceph-mon[57405]: osdmap e99: 8 total, 8 up, 8 in 2026-03-10T11:34:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:10 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:11.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:11 vm09 ceph-mon[54793]: osdmap e100: 8 total, 8 up, 8 in 2026-03-10T11:34:11.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:11 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3625272014' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:11 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:11 vm09 ceph-mon[54793]: pgmap v99: 196 pgs: 31 unknown, 165 active+clean; 455 KiB data, 276 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:11 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:11 vm09 ceph-mon[54793]: osdmap e101: 8 total, 8 up, 8 in 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[49228]: osdmap e100: 8 total, 8 up, 8 in 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3625272014' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[49228]: pgmap v99: 196 pgs: 31 unknown, 165 active+clean; 455 KiB data, 276 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[49228]: osdmap e101: 8 total, 8 up, 8 in 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[57405]: osdmap e100: 8 total, 8 up, 8 in 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3625272014' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[57405]: pgmap v99: 196 pgs: 31 unknown, 165 active+clean; 455 KiB data, 276 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:11 vm06 ceph-mon[57405]: osdmap e101: 8 total, 8 up, 8 in 2026-03-10T11:34:12.729 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_writesame PASSED [ 27%] 2026-03-10T11:34:13.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:13 vm09 ceph-mon[54793]: osdmap e102: 8 total, 8 up, 8 in 2026-03-10T11:34:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:13 vm09 ceph-mon[54793]: pgmap v102: 164 pgs: 164 active+clean; 455 KiB data, 316 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:14.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:13 vm06 ceph-mon[49228]: osdmap e102: 8 total, 8 up, 8 in 2026-03-10T11:34:14.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:13 vm06 ceph-mon[49228]: pgmap v102: 164 pgs: 164 active+clean; 455 KiB data, 316 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:14.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:13 vm06 ceph-mon[57405]: osdmap e102: 8 total, 8 up, 8 in 2026-03-10T11:34:14.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:13 vm06 ceph-mon[57405]: pgmap v102: 164 pgs: 164 active+clean; 455 KiB data, 316 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:14.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:14 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:14.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:14 vm09 ceph-mon[54793]: osdmap e103: 8 total, 8 up, 8 in 2026-03-10T11:34:15.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:14 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:15.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:14 vm06 ceph-mon[49228]: osdmap e103: 8 total, 8 up, 8 in 2026-03-10T11:34:15.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:14 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:15.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:14 vm06 ceph-mon[57405]: osdmap e103: 8 total, 8 up, 8 in 2026-03-10T11:34:16.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:15 vm09 ceph-mon[54793]: osdmap e104: 8 total, 8 up, 8 in 2026-03-10T11:34:16.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:15 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2136502089' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:16.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:15 vm09 ceph-mon[54793]: pgmap v105: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 316 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:16.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:34:16 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:34:16.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:15 vm06 ceph-mon[49228]: osdmap e104: 8 total, 8 up, 8 in 2026-03-10T11:34:16.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:15 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2136502089' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:16.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:15 vm06 ceph-mon[49228]: pgmap v105: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 316 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:16.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:15 vm06 ceph-mon[57405]: osdmap e104: 8 total, 8 up, 8 in 2026-03-10T11:34:16.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:15 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2136502089' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:16.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:15 vm06 ceph-mon[57405]: pgmap v105: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 316 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:17.270 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_append PASSED [ 28%] 2026-03-10T11:34:17.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:17 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2136502089' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:17.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:17 vm06 ceph-mon[49228]: osdmap e105: 8 total, 8 up, 8 in 2026-03-10T11:34:17.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:17 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:17.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:17 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2136502089' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:17.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:17 vm06 ceph-mon[57405]: osdmap e105: 8 total, 8 up, 8 in 2026-03-10T11:34:17.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:17 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:17.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:17 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2136502089' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:17.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:17 vm09 ceph-mon[54793]: osdmap e105: 8 total, 8 up, 8 in 2026-03-10T11:34:17.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:17 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:18.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:18 vm09 ceph-mon[54793]: pgmap v107: 196 pgs: 196 active+clean; 455 KiB data, 338 MiB used, 160 GiB / 160 GiB avail; 1.4 KiB/s rd, 712 B/s wr, 2 op/s 2026-03-10T11:34:18.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:18 vm09 ceph-mon[54793]: osdmap e106: 8 total, 8 up, 8 in 2026-03-10T11:34:18.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:18 vm06 ceph-mon[49228]: pgmap v107: 196 pgs: 196 active+clean; 455 KiB data, 338 MiB used, 160 GiB / 160 GiB avail; 1.4 KiB/s rd, 712 B/s wr, 2 op/s 2026-03-10T11:34:18.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:18 vm06 ceph-mon[49228]: osdmap e106: 8 total, 8 up, 8 in 2026-03-10T11:34:18.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:18 vm06 ceph-mon[57405]: pgmap v107: 196 pgs: 196 active+clean; 455 KiB data, 338 MiB used, 160 GiB / 160 GiB avail; 1.4 KiB/s rd, 712 B/s wr, 2 op/s 2026-03-10T11:34:18.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:18 vm06 ceph-mon[57405]: osdmap e106: 8 total, 8 up, 8 in 2026-03-10T11:34:19.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:19 vm09 ceph-mon[54793]: osdmap e107: 8 total, 8 up, 8 in 2026-03-10T11:34:19.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:19 vm06 ceph-mon[49228]: osdmap e107: 8 total, 8 up, 8 in 2026-03-10T11:34:19.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:19 vm06 ceph-mon[57405]: osdmap e107: 8 total, 8 up, 8 in 2026-03-10T11:34:20.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:20 vm09 ceph-mon[54793]: pgmap v110: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 338 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:20.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:20 vm09 ceph-mon[54793]: osdmap e108: 8 total, 8 up, 8 in 2026-03-10T11:34:20.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:20 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2963157092' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:20.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:20 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:20.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:20 vm06 ceph-mon[49228]: pgmap v110: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 338 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:20.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:20 vm06 ceph-mon[49228]: osdmap e108: 8 total, 8 up, 8 in 2026-03-10T11:34:20.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:20 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2963157092' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:20.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:20 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:20.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:20 vm06 ceph-mon[57405]: pgmap v110: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 338 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:20.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:20 vm06 ceph-mon[57405]: osdmap e108: 8 total, 8 up, 8 in 2026-03-10T11:34:20.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2963157092' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:20.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:20 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:21.030 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:34:20 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:34:20] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:34:21.374 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_write_zeros PASSED [ 29%] 2026-03-10T11:34:21.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:21 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:21.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:21 vm09 ceph-mon[54793]: osdmap e109: 8 total, 8 up, 8 in 2026-03-10T11:34:21.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:21 vm09 ceph-mon[54793]: pgmap v113: 196 pgs: 23 unknown, 173 active+clean; 455 KiB data, 338 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:21.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:21 vm09 ceph-mon[54793]: osdmap e110: 8 total, 8 up, 8 in 2026-03-10T11:34:21.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:21 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:21.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:21 vm06 ceph-mon[49228]: osdmap e109: 8 total, 8 up, 8 in 2026-03-10T11:34:21.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:21 vm06 ceph-mon[49228]: pgmap v113: 196 pgs: 23 unknown, 173 active+clean; 455 KiB data, 338 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:21.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:21 vm06 ceph-mon[49228]: osdmap e110: 8 total, 8 up, 8 in 2026-03-10T11:34:21.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:21 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:21.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:21 vm06 ceph-mon[57405]: osdmap e109: 8 total, 8 up, 8 in 2026-03-10T11:34:21.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:21 vm06 ceph-mon[57405]: pgmap v113: 196 pgs: 23 unknown, 173 active+clean; 455 KiB data, 338 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:21.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:21 vm06 ceph-mon[57405]: osdmap e110: 8 total, 8 up, 8 in 2026-03-10T11:34:23.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:23 vm09 ceph-mon[54793]: osdmap e111: 8 total, 8 up, 8 in 2026-03-10T11:34:23.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:23 vm09 ceph-mon[54793]: pgmap v116: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 342 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:23.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:23 vm06 ceph-mon[49228]: osdmap e111: 8 total, 8 up, 8 in 2026-03-10T11:34:23.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:23 vm06 ceph-mon[49228]: pgmap v116: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 342 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:23.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:23 vm06 ceph-mon[57405]: osdmap e111: 8 total, 8 up, 8 in 2026-03-10T11:34:23.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:23 vm06 ceph-mon[57405]: pgmap v116: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 342 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:24.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:24 vm09 ceph-mon[54793]: osdmap e112: 8 total, 8 up, 8 in 2026-03-10T11:34:24.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:24 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:24.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:24 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2778663574' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:24.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:24 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:24.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:24 vm06 ceph-mon[49228]: osdmap e112: 8 total, 8 up, 8 in 2026-03-10T11:34:24.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:24 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:24.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:24 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2778663574' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:24.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:24 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:24.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:24 vm06 ceph-mon[57405]: osdmap e112: 8 total, 8 up, 8 in 2026-03-10T11:34:24.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:24 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:24.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:24 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2778663574' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:24.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:24 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:25.483 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_trunc PASSED [ 30%] 2026-03-10T11:34:25.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:25 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:25.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:25 vm09 ceph-mon[54793]: osdmap e113: 8 total, 8 up, 8 in 2026-03-10T11:34:25.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:25 vm09 ceph-mon[54793]: pgmap v119: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 342 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:25.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:25.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:25 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:25.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:25 vm06 ceph-mon[49228]: osdmap e113: 8 total, 8 up, 8 in 2026-03-10T11:34:25.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:25 vm06 ceph-mon[49228]: pgmap v119: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 342 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:25.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:25.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:25 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:25.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:25 vm06 ceph-mon[57405]: osdmap e113: 8 total, 8 up, 8 in 2026-03-10T11:34:25.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:25 vm06 ceph-mon[57405]: pgmap v119: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 342 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:25.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:26.474 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:34:26 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:34:26.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:26 vm09 ceph-mon[54793]: osdmap e114: 8 total, 8 up, 8 in 2026-03-10T11:34:26.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:26 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:26.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:26 vm06 ceph-mon[49228]: osdmap e114: 8 total, 8 up, 8 in 2026-03-10T11:34:26.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:26 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:26.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:26 vm06 ceph-mon[57405]: osdmap e114: 8 total, 8 up, 8 in 2026-03-10T11:34:26.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:26 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:27.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:27 vm06 ceph-mon[49228]: osdmap e115: 8 total, 8 up, 8 in 2026-03-10T11:34:27.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:27 vm06 ceph-mon[49228]: pgmap v122: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:27.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:27 vm06 ceph-mon[57405]: osdmap e115: 8 total, 8 up, 8 in 2026-03-10T11:34:27.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:27 vm06 ceph-mon[57405]: pgmap v122: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:27.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:27 vm09 ceph-mon[54793]: osdmap e115: 8 total, 8 up, 8 in 2026-03-10T11:34:27.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:27 vm09 ceph-mon[54793]: pgmap v122: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:28.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:28 vm06 ceph-mon[49228]: osdmap e116: 8 total, 8 up, 8 in 2026-03-10T11:34:28.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:28 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2057623633' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:28.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:28 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:28.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:28 vm06 ceph-mon[57405]: osdmap e116: 8 total, 8 up, 8 in 2026-03-10T11:34:28.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:28 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2057623633' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:28.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:28 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:28.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:28 vm09 ceph-mon[54793]: osdmap e116: 8 total, 8 up, 8 in 2026-03-10T11:34:28.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:28 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2057623633' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:28.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:28 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:29.499 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_cmpext PASSED [ 31%] 2026-03-10T11:34:29.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:29 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:29.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:29 vm06 ceph-mon[49228]: osdmap e117: 8 total, 8 up, 8 in 2026-03-10T11:34:29.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:29 vm06 ceph-mon[49228]: pgmap v125: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:29.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:29 vm06 ceph-mon[49228]: osdmap e118: 8 total, 8 up, 8 in 2026-03-10T11:34:29.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:29 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:29.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:29 vm06 ceph-mon[57405]: osdmap e117: 8 total, 8 up, 8 in 2026-03-10T11:34:29.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:29 vm06 ceph-mon[57405]: pgmap v125: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:29.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:29 vm06 ceph-mon[57405]: osdmap e118: 8 total, 8 up, 8 in 2026-03-10T11:34:29.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:29 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:29.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:29 vm09 ceph-mon[54793]: osdmap e117: 8 total, 8 up, 8 in 2026-03-10T11:34:29.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:29 vm09 ceph-mon[54793]: pgmap v125: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:29.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:29 vm09 ceph-mon[54793]: osdmap e118: 8 total, 8 up, 8 in 2026-03-10T11:34:31.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:34:30 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:34:30] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:34:31.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:31 vm09 ceph-mon[54793]: osdmap e119: 8 total, 8 up, 8 in 2026-03-10T11:34:31.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:31 vm09 ceph-mon[54793]: pgmap v128: 196 pgs: 11 creating+peering, 21 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:32.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:31 vm06 ceph-mon[49228]: osdmap e119: 8 total, 8 up, 8 in 2026-03-10T11:34:32.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:31 vm06 ceph-mon[49228]: pgmap v128: 196 pgs: 11 creating+peering, 21 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:32.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:31 vm06 ceph-mon[57405]: osdmap e119: 8 total, 8 up, 8 in 2026-03-10T11:34:32.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:31 vm06 ceph-mon[57405]: pgmap v128: 196 pgs: 11 creating+peering, 21 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:32.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:32 vm09 ceph-mon[54793]: osdmap e120: 8 total, 8 up, 8 in 2026-03-10T11:34:32.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:32 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2175853670' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:32.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:32 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:32.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:32 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:32.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:32 vm09 ceph-mon[54793]: osdmap e121: 8 total, 8 up, 8 in 2026-03-10T11:34:33.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:32 vm06 ceph-mon[49228]: osdmap e120: 8 total, 8 up, 8 in 2026-03-10T11:34:33.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:32 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2175853670' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:33.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:32 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:33.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:32 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:33.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:32 vm06 ceph-mon[49228]: osdmap e121: 8 total, 8 up, 8 in 2026-03-10T11:34:33.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:32 vm06 ceph-mon[57405]: osdmap e120: 8 total, 8 up, 8 in 2026-03-10T11:34:33.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:32 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2175853670' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:33.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:32 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:33.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:32 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:33.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:32 vm06 ceph-mon[57405]: osdmap e121: 8 total, 8 up, 8 in 2026-03-10T11:34:33.537 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_list_objects_empty PASSED [ 32%] 2026-03-10T11:34:33.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:33 vm09 ceph-mon[54793]: pgmap v131: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:33.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:33 vm09 ceph-mon[54793]: osdmap e122: 8 total, 8 up, 8 in 2026-03-10T11:34:34.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:33 vm06 ceph-mon[49228]: pgmap v131: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:34.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:33 vm06 ceph-mon[49228]: osdmap e122: 8 total, 8 up, 8 in 2026-03-10T11:34:34.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:33 vm06 ceph-mon[57405]: pgmap v131: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:34.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:33 vm06 ceph-mon[57405]: osdmap e122: 8 total, 8 up, 8 in 2026-03-10T11:34:35.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:35 vm09 ceph-mon[54793]: osdmap e123: 8 total, 8 up, 8 in 2026-03-10T11:34:35.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:35 vm09 ceph-mon[54793]: pgmap v134: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:35 vm06 ceph-mon[49228]: osdmap e123: 8 total, 8 up, 8 in 2026-03-10T11:34:36.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:35 vm06 ceph-mon[49228]: pgmap v134: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:35 vm06 ceph-mon[57405]: osdmap e123: 8 total, 8 up, 8 in 2026-03-10T11:34:36.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:35 vm06 ceph-mon[57405]: pgmap v134: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 343 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:36.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:34:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:34:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:36 vm09 ceph-mon[54793]: osdmap e124: 8 total, 8 up, 8 in 2026-03-10T11:34:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:36 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3211157325' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:36 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:36.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:36 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:37.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:36 vm06 ceph-mon[49228]: osdmap e124: 8 total, 8 up, 8 in 2026-03-10T11:34:37.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:36 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3211157325' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:37.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:36 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:37.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:36 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:37.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:36 vm06 ceph-mon[57405]: osdmap e124: 8 total, 8 up, 8 in 2026-03-10T11:34:37.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:36 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3211157325' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:37.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:36 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:37.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:36 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:37.609 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_read_crc PASSED [ 34%] 2026-03-10T11:34:37.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:37 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:37.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:37 vm09 ceph-mon[54793]: osdmap e125: 8 total, 8 up, 8 in 2026-03-10T11:34:37.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:37 vm09 ceph-mon[54793]: pgmap v137: 196 pgs: 196 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 0 B/s wr, 1 op/s 2026-03-10T11:34:38.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:37 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:38.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:37 vm06 ceph-mon[49228]: osdmap e125: 8 total, 8 up, 8 in 2026-03-10T11:34:38.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:37 vm06 ceph-mon[49228]: pgmap v137: 196 pgs: 196 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 0 B/s wr, 1 op/s 2026-03-10T11:34:38.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:37 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:38.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:37 vm06 ceph-mon[57405]: osdmap e125: 8 total, 8 up, 8 in 2026-03-10T11:34:38.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:37 vm06 ceph-mon[57405]: pgmap v137: 196 pgs: 196 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 0 B/s wr, 1 op/s 2026-03-10T11:34:38.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:38 vm09 ceph-mon[54793]: osdmap e126: 8 total, 8 up, 8 in 2026-03-10T11:34:39.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:38 vm06 ceph-mon[49228]: osdmap e126: 8 total, 8 up, 8 in 2026-03-10T11:34:39.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:38 vm06 ceph-mon[57405]: osdmap e126: 8 total, 8 up, 8 in 2026-03-10T11:34:39.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:39 vm09 ceph-mon[54793]: osdmap e127: 8 total, 8 up, 8 in 2026-03-10T11:34:39.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:39 vm09 ceph-mon[54793]: pgmap v140: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:40.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:39 vm06 ceph-mon[49228]: osdmap e127: 8 total, 8 up, 8 in 2026-03-10T11:34:40.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:39 vm06 ceph-mon[49228]: pgmap v140: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:40.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:39 vm06 ceph-mon[57405]: osdmap e127: 8 total, 8 up, 8 in 2026-03-10T11:34:40.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:39 vm06 ceph-mon[57405]: pgmap v140: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:40.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:40 vm09 ceph-mon[54793]: osdmap e128: 8 total, 8 up, 8 in 2026-03-10T11:34:40.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:40 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/4002684719' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:40.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:40 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:40.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:40 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:41.030 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:34:40 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:34:40] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:34:41.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:40 vm06 ceph-mon[49228]: osdmap e128: 8 total, 8 up, 8 in 2026-03-10T11:34:41.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:40 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/4002684719' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:41.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:40 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:41.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:40 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:41.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:40 vm06 ceph-mon[57405]: osdmap e128: 8 total, 8 up, 8 in 2026-03-10T11:34:41.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:40 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/4002684719' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:41.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:40 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:41.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:40 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:41.654 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_list_objects PASSED [ 35%] 2026-03-10T11:34:41.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:41 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:41.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:41 vm09 ceph-mon[54793]: osdmap e129: 8 total, 8 up, 8 in 2026-03-10T11:34:41.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:41 vm09 ceph-mon[54793]: pgmap v143: 196 pgs: 26 unknown, 170 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:42.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:41 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:42.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:41 vm06 ceph-mon[49228]: osdmap e129: 8 total, 8 up, 8 in 2026-03-10T11:34:42.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:41 vm06 ceph-mon[49228]: pgmap v143: 196 pgs: 26 unknown, 170 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:42.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:41 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:42.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:41 vm06 ceph-mon[57405]: osdmap e129: 8 total, 8 up, 8 in 2026-03-10T11:34:42.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:41 vm06 ceph-mon[57405]: pgmap v143: 196 pgs: 26 unknown, 170 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:42.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:42 vm09 ceph-mon[54793]: osdmap e130: 8 total, 8 up, 8 in 2026-03-10T11:34:43.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:42 vm06 ceph-mon[49228]: osdmap e130: 8 total, 8 up, 8 in 2026-03-10T11:34:43.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:42 vm06 ceph-mon[57405]: osdmap e130: 8 total, 8 up, 8 in 2026-03-10T11:34:43.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:43 vm09 ceph-mon[54793]: osdmap e131: 8 total, 8 up, 8 in 2026-03-10T11:34:43.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:43 vm09 ceph-mon[54793]: pgmap v146: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:44.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:43 vm06 ceph-mon[49228]: osdmap e131: 8 total, 8 up, 8 in 2026-03-10T11:34:44.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:43 vm06 ceph-mon[49228]: pgmap v146: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:44.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:43 vm06 ceph-mon[57405]: osdmap e131: 8 total, 8 up, 8 in 2026-03-10T11:34:44.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:43 vm06 ceph-mon[57405]: pgmap v146: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:44.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:44 vm09 ceph-mon[54793]: osdmap e132: 8 total, 8 up, 8 in 2026-03-10T11:34:44.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:44 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3643794371' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:44.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:44 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:45.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:44 vm06 ceph-mon[49228]: osdmap e132: 8 total, 8 up, 8 in 2026-03-10T11:34:45.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:44 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3643794371' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:45.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:44 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:45.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:44 vm06 ceph-mon[57405]: osdmap e132: 8 total, 8 up, 8 in 2026-03-10T11:34:45.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:44 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3643794371' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:45.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:44 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:45.706 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_list_ns_objects PASSED [ 36%] 2026-03-10T11:34:45.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:45 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:45.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:45 vm09 ceph-mon[54793]: osdmap e133: 8 total, 8 up, 8 in 2026-03-10T11:34:45.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:45 vm09 ceph-mon[54793]: pgmap v149: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:46.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:45 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:46.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:45 vm06 ceph-mon[49228]: osdmap e133: 8 total, 8 up, 8 in 2026-03-10T11:34:46.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:45 vm06 ceph-mon[49228]: pgmap v149: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:46.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:45 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:46.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:45 vm06 ceph-mon[57405]: osdmap e133: 8 total, 8 up, 8 in 2026-03-10T11:34:46.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:45 vm06 ceph-mon[57405]: pgmap v149: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 344 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:46.479 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:34:46 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:34:46.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:46 vm09 ceph-mon[54793]: osdmap e134: 8 total, 8 up, 8 in 2026-03-10T11:34:46.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:46 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:47.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:46 vm06 ceph-mon[49228]: osdmap e134: 8 total, 8 up, 8 in 2026-03-10T11:34:47.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:46 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:47.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:46 vm06 ceph-mon[57405]: osdmap e134: 8 total, 8 up, 8 in 2026-03-10T11:34:47.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:46 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:47.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:47 vm09 ceph-mon[54793]: osdmap e135: 8 total, 8 up, 8 in 2026-03-10T11:34:47.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:47 vm09 ceph-mon[54793]: pgmap v152: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:48.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:47 vm06 ceph-mon[49228]: osdmap e135: 8 total, 8 up, 8 in 2026-03-10T11:34:48.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:47 vm06 ceph-mon[49228]: pgmap v152: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:47 vm06 ceph-mon[57405]: osdmap e135: 8 total, 8 up, 8 in 2026-03-10T11:34:48.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:47 vm06 ceph-mon[57405]: pgmap v152: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:48.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:48 vm09 ceph-mon[54793]: osdmap e136: 8 total, 8 up, 8 in 2026-03-10T11:34:48.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:48 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2058782250' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:49.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:48 vm06 ceph-mon[49228]: osdmap e136: 8 total, 8 up, 8 in 2026-03-10T11:34:49.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:48 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2058782250' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:49.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:48 vm06 ceph-mon[57405]: osdmap e136: 8 total, 8 up, 8 in 2026-03-10T11:34:49.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:48 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2058782250' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:49.745 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_xattrs PASSED [ 37%] 2026-03-10T11:34:50.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:49 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2058782250' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:49 vm06 ceph-mon[49228]: osdmap e137: 8 total, 8 up, 8 in 2026-03-10T11:34:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:49 vm06 ceph-mon[49228]: pgmap v155: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:49 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2058782250' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:49 vm06 ceph-mon[57405]: osdmap e137: 8 total, 8 up, 8 in 2026-03-10T11:34:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:49 vm06 ceph-mon[57405]: pgmap v155: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:50.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:49 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2058782250' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:50.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:49 vm09 ceph-mon[54793]: osdmap e137: 8 total, 8 up, 8 in 2026-03-10T11:34:50.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:49 vm09 ceph-mon[54793]: pgmap v155: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:51.030 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:34:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:34:50] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:34:51.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:50 vm06 ceph-mon[49228]: osdmap e138: 8 total, 8 up, 8 in 2026-03-10T11:34:51.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:50 vm06 ceph-mon[57405]: osdmap e138: 8 total, 8 up, 8 in 2026-03-10T11:34:51.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:50 vm09 ceph-mon[54793]: osdmap e138: 8 total, 8 up, 8 in 2026-03-10T11:34:52.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:51 vm06 ceph-mon[49228]: osdmap e139: 8 total, 8 up, 8 in 2026-03-10T11:34:52.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:51 vm06 ceph-mon[49228]: pgmap v158: 196 pgs: 13 creating+peering, 19 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:52.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:51 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:34:52.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:51 vm06 ceph-mon[57405]: osdmap e139: 8 total, 8 up, 8 in 2026-03-10T11:34:52.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:51 vm06 ceph-mon[57405]: pgmap v158: 196 pgs: 13 creating+peering, 19 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:52.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:51 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:34:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:51 vm09 ceph-mon[54793]: osdmap e139: 8 total, 8 up, 8 in 2026-03-10T11:34:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:51 vm09 ceph-mon[54793]: pgmap v158: 196 pgs: 13 creating+peering, 19 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:34:52.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:51 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:34:53.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[49228]: osdmap e140: 8 total, 8 up, 8 in 2026-03-10T11:34:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:34:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:34:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:34:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3500049451' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[57405]: osdmap e140: 8 total, 8 up, 8 in 2026-03-10T11:34:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:34:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:34:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:34:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3500049451' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:52 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:53.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:52 vm09 ceph-mon[54793]: osdmap e140: 8 total, 8 up, 8 in 2026-03-10T11:34:53.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:52 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:34:53.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:52 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:34:53.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:52 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:34:53.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:52 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3500049451' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:53.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:52 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:53.813 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_obj_xattrs PASSED [ 38%] 2026-03-10T11:34:54.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:53 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:54.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:53 vm09 ceph-mon[54793]: osdmap e141: 8 total, 8 up, 8 in 2026-03-10T11:34:54.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:53 vm09 ceph-mon[54793]: pgmap v161: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:54.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:53 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:54.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:53 vm06 ceph-mon[49228]: osdmap e141: 8 total, 8 up, 8 in 2026-03-10T11:34:54.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:53 vm06 ceph-mon[49228]: pgmap v161: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:54.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:53 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:54.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:53 vm06 ceph-mon[57405]: osdmap e141: 8 total, 8 up, 8 in 2026-03-10T11:34:54.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:53 vm06 ceph-mon[57405]: pgmap v161: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:55.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:54 vm09 ceph-mon[54793]: osdmap e142: 8 total, 8 up, 8 in 2026-03-10T11:34:55.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:54 vm06 ceph-mon[49228]: osdmap e142: 8 total, 8 up, 8 in 2026-03-10T11:34:55.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:54 vm06 ceph-mon[57405]: osdmap e142: 8 total, 8 up, 8 in 2026-03-10T11:34:56.164 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:55 vm09 ceph-mon[54793]: osdmap e143: 8 total, 8 up, 8 in 2026-03-10T11:34:56.164 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:55 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3467437774' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:56.164 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:55 vm09 ceph-mon[54793]: pgmap v164: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:56.164 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:55 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:56.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:55 vm06 ceph-mon[49228]: osdmap e143: 8 total, 8 up, 8 in 2026-03-10T11:34:56.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:55 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3467437774' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:56.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:55 vm06 ceph-mon[49228]: pgmap v164: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:56.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:55 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:56.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:55 vm06 ceph-mon[57405]: osdmap e143: 8 total, 8 up, 8 in 2026-03-10T11:34:56.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:55 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3467437774' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:56.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:55 vm06 ceph-mon[57405]: pgmap v164: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 345 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:56.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:55 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:34:56.479 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:34:56 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:34:56.825 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_get_pool_id PASSED [ 39%] 2026-03-10T11:34:57.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:56 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3467437774' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:57.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:56 vm09 ceph-mon[54793]: osdmap e144: 8 total, 8 up, 8 in 2026-03-10T11:34:57.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:56 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:57.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:56 vm09 ceph-mon[54793]: osdmap e145: 8 total, 8 up, 8 in 2026-03-10T11:34:57.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:56 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3467437774' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:57.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:56 vm06 ceph-mon[49228]: osdmap e144: 8 total, 8 up, 8 in 2026-03-10T11:34:57.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:56 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:57.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:56 vm06 ceph-mon[49228]: osdmap e145: 8 total, 8 up, 8 in 2026-03-10T11:34:57.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:56 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3467437774' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:57.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:56 vm06 ceph-mon[57405]: osdmap e144: 8 total, 8 up, 8 in 2026-03-10T11:34:57.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:56 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:34:57.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:56 vm06 ceph-mon[57405]: osdmap e145: 8 total, 8 up, 8 in 2026-03-10T11:34:58.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:57 vm09 ceph-mon[54793]: pgmap v167: 164 pgs: 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:58.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:57 vm09 ceph-mon[54793]: osdmap e146: 8 total, 8 up, 8 in 2026-03-10T11:34:58.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:57 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/226796886' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:58.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:57 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:58.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:57 vm06 ceph-mon[49228]: pgmap v167: 164 pgs: 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:57 vm06 ceph-mon[49228]: osdmap e146: 8 total, 8 up, 8 in 2026-03-10T11:34:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:57 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/226796886' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:58.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:57 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:58.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:57 vm06 ceph-mon[57405]: pgmap v167: 164 pgs: 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:34:58.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:57 vm06 ceph-mon[57405]: osdmap e146: 8 total, 8 up, 8 in 2026-03-10T11:34:58.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:57 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/226796886' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:58.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:57 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:34:59.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:58 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:59.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:58 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:59.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:58 vm09 ceph-mon[54793]: osdmap e147: 8 total, 8 up, 8 in 2026-03-10T11:34:59.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:58 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:59.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:58 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:59.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:58 vm06 ceph-mon[49228]: osdmap e147: 8 total, 8 up, 8 in 2026-03-10T11:34:59.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:58 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:34:59.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:58 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:34:59.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:58 vm06 ceph-mon[57405]: osdmap e147: 8 total, 8 up, 8 in 2026-03-10T11:34:59.870 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_get_pool_name PASSED [ 40%] 2026-03-10T11:35:00.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:34:59 vm09 ceph-mon[54793]: pgmap v170: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:00.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:34:59 vm06 ceph-mon[49228]: pgmap v170: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:00.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:34:59 vm06 ceph-mon[57405]: pgmap v170: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:01.030 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:00 vm06 ceph-mon[57405]: osdmap e148: 8 total, 8 up, 8 in 2026-03-10T11:35:01.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:35:00 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:35:00] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:35:01.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:00 vm06 ceph-mon[49228]: osdmap e148: 8 total, 8 up, 8 in 2026-03-10T11:35:01.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:00 vm09 ceph-mon[54793]: osdmap e148: 8 total, 8 up, 8 in 2026-03-10T11:35:02.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:01 vm09 ceph-mon[54793]: osdmap e149: 8 total, 8 up, 8 in 2026-03-10T11:35:02.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:01 vm09 ceph-mon[54793]: pgmap v173: 196 pgs: 10 creating+peering, 22 unknown, 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:02.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:01 vm09 ceph-mon[54793]: osdmap e150: 8 total, 8 up, 8 in 2026-03-10T11:35:02.280 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:01 vm06 ceph-mon[57405]: osdmap e149: 8 total, 8 up, 8 in 2026-03-10T11:35:02.280 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:01 vm06 ceph-mon[57405]: pgmap v173: 196 pgs: 10 creating+peering, 22 unknown, 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:02.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:01 vm06 ceph-mon[57405]: osdmap e150: 8 total, 8 up, 8 in 2026-03-10T11:35:02.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:01 vm06 ceph-mon[49228]: osdmap e149: 8 total, 8 up, 8 in 2026-03-10T11:35:02.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:01 vm06 ceph-mon[49228]: pgmap v173: 196 pgs: 10 creating+peering, 22 unknown, 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:02.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:01 vm06 ceph-mon[49228]: osdmap e150: 8 total, 8 up, 8 in 2026-03-10T11:35:04.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:03 vm09 ceph-mon[54793]: osdmap e151: 8 total, 8 up, 8 in 2026-03-10T11:35:04.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:03 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/658310154' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:04.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:03 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:04.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:03 vm09 ceph-mon[54793]: pgmap v176: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:04.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:03 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:35:04.280 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:03 vm06 ceph-mon[57405]: osdmap e151: 8 total, 8 up, 8 in 2026-03-10T11:35:04.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:03 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/658310154' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:04.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:03 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:04.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:03 vm06 ceph-mon[57405]: pgmap v176: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:04.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:03 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:35:04.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:03 vm06 ceph-mon[49228]: osdmap e151: 8 total, 8 up, 8 in 2026-03-10T11:35:04.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:03 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/658310154' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:04.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:03 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:04.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:03 vm06 ceph-mon[49228]: pgmap v176: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:04.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:03 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:35:04.953 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_create_snap PASSED [ 41%] 2026-03-10T11:35:05.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:04 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:04 vm09 ceph-mon[54793]: osdmap e152: 8 total, 8 up, 8 in 2026-03-10T11:35:05.280 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:04 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:05.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:04 vm06 ceph-mon[57405]: osdmap e152: 8 total, 8 up, 8 in 2026-03-10T11:35:05.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:04 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:05.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:04 vm06 ceph-mon[49228]: osdmap e152: 8 total, 8 up, 8 in 2026-03-10T11:35:06.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:05 vm09 ceph-mon[54793]: osdmap e153: 8 total, 8 up, 8 in 2026-03-10T11:35:06.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:05 vm09 ceph-mon[54793]: pgmap v179: 164 pgs: 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:06.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:35:06 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:35:06.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:05 vm06 ceph-mon[49228]: osdmap e153: 8 total, 8 up, 8 in 2026-03-10T11:35:06.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:05 vm06 ceph-mon[49228]: pgmap v179: 164 pgs: 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:06.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:05 vm06 ceph-mon[57405]: osdmap e153: 8 total, 8 up, 8 in 2026-03-10T11:35:06.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:05 vm06 ceph-mon[57405]: pgmap v179: 164 pgs: 164 active+clean; 455 KiB data, 346 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:07.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:06 vm06 ceph-mon[49228]: osdmap e154: 8 total, 8 up, 8 in 2026-03-10T11:35:07.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:06 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1301227451' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:07.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:06 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:07.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:06 vm06 ceph-mon[57405]: osdmap e154: 8 total, 8 up, 8 in 2026-03-10T11:35:07.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:06 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1301227451' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:07.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:06 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:07.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:06 vm09 ceph-mon[54793]: osdmap e154: 8 total, 8 up, 8 in 2026-03-10T11:35:07.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:06 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1301227451' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:07.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:06 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:07.968 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_list_snaps_empty PASSED [ 42%] 2026-03-10T11:35:08.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:07 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1301227451' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:08.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:07 vm06 ceph-mon[49228]: osdmap e155: 8 total, 8 up, 8 in 2026-03-10T11:35:08.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:07 vm06 ceph-mon[49228]: pgmap v182: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:08.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:07 vm06 ceph-mon[49228]: osdmap e156: 8 total, 8 up, 8 in 2026-03-10T11:35:08.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:07 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1301227451' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:08.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:07 vm06 ceph-mon[57405]: osdmap e155: 8 total, 8 up, 8 in 2026-03-10T11:35:08.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:07 vm06 ceph-mon[57405]: pgmap v182: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:08.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:07 vm06 ceph-mon[57405]: osdmap e156: 8 total, 8 up, 8 in 2026-03-10T11:35:08.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:07 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1301227451' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:08.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:07 vm09 ceph-mon[54793]: osdmap e155: 8 total, 8 up, 8 in 2026-03-10T11:35:08.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:07 vm09 ceph-mon[54793]: pgmap v182: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:08.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:07 vm09 ceph-mon[54793]: osdmap e156: 8 total, 8 up, 8 in 2026-03-10T11:35:10.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:10 vm06 ceph-mon[49228]: osdmap e157: 8 total, 8 up, 8 in 2026-03-10T11:35:10.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:10 vm06 ceph-mon[49228]: pgmap v185: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:10.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:10 vm06 ceph-mon[57405]: osdmap e157: 8 total, 8 up, 8 in 2026-03-10T11:35:10.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:10 vm06 ceph-mon[57405]: pgmap v185: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:10.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:10 vm09 ceph-mon[54793]: osdmap e157: 8 total, 8 up, 8 in 2026-03-10T11:35:10.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:10 vm09 ceph-mon[54793]: pgmap v185: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:11.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:35:10 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:35:10] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:35:11.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:11 vm09 ceph-mon[54793]: osdmap e158: 8 total, 8 up, 8 in 2026-03-10T11:35:11.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:11 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:11.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:11 vm06 ceph-mon[49228]: osdmap e158: 8 total, 8 up, 8 in 2026-03-10T11:35:11.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:11 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:11.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:11 vm06 ceph-mon[57405]: osdmap e158: 8 total, 8 up, 8 in 2026-03-10T11:35:11.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:11 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:12.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:12 vm09 ceph-mon[54793]: osdmap e159: 8 total, 8 up, 8 in 2026-03-10T11:35:12.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:12 vm09 ceph-mon[54793]: pgmap v188: 196 pgs: 20 unknown, 176 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:12.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:12 vm09 ceph-mon[54793]: osdmap e160: 8 total, 8 up, 8 in 2026-03-10T11:35:12.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:12 vm06 ceph-mon[49228]: osdmap e159: 8 total, 8 up, 8 in 2026-03-10T11:35:12.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:12 vm06 ceph-mon[49228]: pgmap v188: 196 pgs: 20 unknown, 176 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:12.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:12 vm06 ceph-mon[49228]: osdmap e160: 8 total, 8 up, 8 in 2026-03-10T11:35:12.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:12 vm06 ceph-mon[57405]: osdmap e159: 8 total, 8 up, 8 in 2026-03-10T11:35:12.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:12 vm06 ceph-mon[57405]: pgmap v188: 196 pgs: 20 unknown, 176 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:12.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:12 vm06 ceph-mon[57405]: osdmap e160: 8 total, 8 up, 8 in 2026-03-10T11:35:13.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:13 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1319300707' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:13.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:13 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:13.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:13 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1319300707' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:13.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:13 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:13.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:13 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1319300707' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:13.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:13 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:14.063 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_list_snaps PASSED [ 43%] 2026-03-10T11:35:14.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:14 vm09 ceph-mon[54793]: pgmap v190: 196 pgs: 196 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:14.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:14 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:14.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:14 vm09 ceph-mon[54793]: osdmap e161: 8 total, 8 up, 8 in 2026-03-10T11:35:14.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:14 vm06 ceph-mon[49228]: pgmap v190: 196 pgs: 196 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:14.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:14 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:14.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:14 vm06 ceph-mon[49228]: osdmap e161: 8 total, 8 up, 8 in 2026-03-10T11:35:14.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:14 vm06 ceph-mon[57405]: pgmap v190: 196 pgs: 196 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:14.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:14 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:14.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:14 vm06 ceph-mon[57405]: osdmap e161: 8 total, 8 up, 8 in 2026-03-10T11:35:15.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:15 vm09 ceph-mon[54793]: osdmap e162: 8 total, 8 up, 8 in 2026-03-10T11:35:15.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:15 vm06 ceph-mon[49228]: osdmap e162: 8 total, 8 up, 8 in 2026-03-10T11:35:15.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:15 vm06 ceph-mon[57405]: osdmap e162: 8 total, 8 up, 8 in 2026-03-10T11:35:16.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:16 vm09 ceph-mon[54793]: pgmap v193: 164 pgs: 164 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:16.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:16 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:35:16.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:16 vm09 ceph-mon[54793]: osdmap e163: 8 total, 8 up, 8 in 2026-03-10T11:35:16.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:35:16 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:35:16.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:16 vm06 ceph-mon[49228]: pgmap v193: 164 pgs: 164 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:16.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:16 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:35:16.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:16 vm06 ceph-mon[49228]: osdmap e163: 8 total, 8 up, 8 in 2026-03-10T11:35:16.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:16 vm06 ceph-mon[57405]: pgmap v193: 164 pgs: 164 active+clean; 455 KiB data, 347 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:16.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:16 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:35:16.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:16 vm06 ceph-mon[57405]: osdmap e163: 8 total, 8 up, 8 in 2026-03-10T11:35:17.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:17 vm09 ceph-mon[54793]: osdmap e164: 8 total, 8 up, 8 in 2026-03-10T11:35:17.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:17 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3733952807' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:17.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:17 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:17.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:17 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:17.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:17 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:17.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:17 vm09 ceph-mon[54793]: osdmap e165: 8 total, 8 up, 8 in 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[49228]: osdmap e164: 8 total, 8 up, 8 in 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3733952807' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[49228]: osdmap e165: 8 total, 8 up, 8 in 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[57405]: osdmap e164: 8 total, 8 up, 8 in 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3733952807' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:17.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:17 vm06 ceph-mon[57405]: osdmap e165: 8 total, 8 up, 8 in 2026-03-10T11:35:18.099 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_lookup_snap PASSED [ 45%] 2026-03-10T11:35:18.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:18 vm09 ceph-mon[54793]: pgmap v196: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:18.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:18 vm09 ceph-mon[54793]: osdmap e166: 8 total, 8 up, 8 in 2026-03-10T11:35:18.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:18 vm06 ceph-mon[49228]: pgmap v196: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:18.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:18 vm06 ceph-mon[49228]: osdmap e166: 8 total, 8 up, 8 in 2026-03-10T11:35:18.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:18 vm06 ceph-mon[57405]: pgmap v196: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:18.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:18 vm06 ceph-mon[57405]: osdmap e166: 8 total, 8 up, 8 in 2026-03-10T11:35:20.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:20 vm09 ceph-mon[54793]: pgmap v199: 164 pgs: 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:20.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:20 vm09 ceph-mon[54793]: osdmap e167: 8 total, 8 up, 8 in 2026-03-10T11:35:20.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:20 vm06 ceph-mon[49228]: pgmap v199: 164 pgs: 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:20.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:20 vm06 ceph-mon[49228]: osdmap e167: 8 total, 8 up, 8 in 2026-03-10T11:35:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:20 vm06 ceph-mon[57405]: pgmap v199: 164 pgs: 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:20.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:20 vm06 ceph-mon[57405]: osdmap e167: 8 total, 8 up, 8 in 2026-03-10T11:35:21.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:35:20 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:35:20] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:35:21.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:21 vm09 ceph-mon[54793]: osdmap e168: 8 total, 8 up, 8 in 2026-03-10T11:35:21.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:21 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/736115250' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:21.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:21 vm06 ceph-mon[49228]: osdmap e168: 8 total, 8 up, 8 in 2026-03-10T11:35:21.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:21 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/736115250' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:21.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:21 vm06 ceph-mon[57405]: osdmap e168: 8 total, 8 up, 8 in 2026-03-10T11:35:21.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:21 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/736115250' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:22.136 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_snap_timestamp PASSED [ 46%] 2026-03-10T11:35:22.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:22 vm09 ceph-mon[54793]: pgmap v202: 196 pgs: 13 unknown, 183 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:22.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:22 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:35:22.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:22 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/736115250' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:22.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:22 vm09 ceph-mon[54793]: osdmap e169: 8 total, 8 up, 8 in 2026-03-10T11:35:22.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:22 vm06 ceph-mon[49228]: pgmap v202: 196 pgs: 13 unknown, 183 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:22.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:22 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:35:22.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:22 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/736115250' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:22.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:22 vm06 ceph-mon[49228]: osdmap e169: 8 total, 8 up, 8 in 2026-03-10T11:35:22.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:22 vm06 ceph-mon[57405]: pgmap v202: 196 pgs: 13 unknown, 183 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:22.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:22 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:35:22.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:22 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/736115250' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:22.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:22 vm06 ceph-mon[57405]: osdmap e169: 8 total, 8 up, 8 in 2026-03-10T11:35:23.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:23 vm09 ceph-mon[54793]: osdmap e170: 8 total, 8 up, 8 in 2026-03-10T11:35:23.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:23 vm06 ceph-mon[49228]: osdmap e170: 8 total, 8 up, 8 in 2026-03-10T11:35:23.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:23 vm06 ceph-mon[57405]: osdmap e170: 8 total, 8 up, 8 in 2026-03-10T11:35:24.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:24 vm09 ceph-mon[54793]: pgmap v205: 164 pgs: 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:24.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:24 vm09 ceph-mon[54793]: osdmap e171: 8 total, 8 up, 8 in 2026-03-10T11:35:24.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:24 vm06 ceph-mon[49228]: pgmap v205: 164 pgs: 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:24.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:24 vm06 ceph-mon[49228]: osdmap e171: 8 total, 8 up, 8 in 2026-03-10T11:35:24.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:24 vm06 ceph-mon[57405]: pgmap v205: 164 pgs: 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:24.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:24 vm06 ceph-mon[57405]: osdmap e171: 8 total, 8 up, 8 in 2026-03-10T11:35:25.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:25 vm09 ceph-mon[54793]: osdmap e172: 8 total, 8 up, 8 in 2026-03-10T11:35:25.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:25 vm06 ceph-mon[49228]: osdmap e172: 8 total, 8 up, 8 in 2026-03-10T11:35:25.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:25 vm06 ceph-mon[57405]: osdmap e172: 8 total, 8 up, 8 in 2026-03-10T11:35:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:26 vm09 ceph-mon[54793]: pgmap v208: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:26 vm09 ceph-mon[54793]: osdmap e173: 8 total, 8 up, 8 in 2026-03-10T11:35:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:26 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/457556044' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:26 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:26 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:26.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:35:26 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:35:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:26 vm06 ceph-mon[49228]: pgmap v208: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:26 vm06 ceph-mon[49228]: osdmap e173: 8 total, 8 up, 8 in 2026-03-10T11:35:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:26 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/457556044' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:26 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:26.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:26 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:26 vm06 ceph-mon[57405]: pgmap v208: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 348 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:26 vm06 ceph-mon[57405]: osdmap e173: 8 total, 8 up, 8 in 2026-03-10T11:35:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:26 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/457556044' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:26 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:26 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:27.219 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_remove_snap PASSED [ 47%] 2026-03-10T11:35:27.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:27 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:27 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:27 vm09 ceph-mon[54793]: osdmap e174: 8 total, 8 up, 8 in 2026-03-10T11:35:27.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:27 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:27.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:27 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:27.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:27 vm06 ceph-mon[49228]: osdmap e174: 8 total, 8 up, 8 in 2026-03-10T11:35:27.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:27 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:27.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:27 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:27.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:27 vm06 ceph-mon[57405]: osdmap e174: 8 total, 8 up, 8 in 2026-03-10T11:35:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:28 vm06 ceph-mon[49228]: pgmap v211: 196 pgs: 196 active+clean; 455 KiB data, 349 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:28.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:28 vm06 ceph-mon[49228]: osdmap e175: 8 total, 8 up, 8 in 2026-03-10T11:35:28.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:28 vm06 ceph-mon[57405]: pgmap v211: 196 pgs: 196 active+clean; 455 KiB data, 349 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:28.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:28 vm06 ceph-mon[57405]: osdmap e175: 8 total, 8 up, 8 in 2026-03-10T11:35:28.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:28 vm09 ceph-mon[54793]: pgmap v211: 196 pgs: 196 active+clean; 455 KiB data, 349 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:28.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:28 vm09 ceph-mon[54793]: osdmap e175: 8 total, 8 up, 8 in 2026-03-10T11:35:29.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:29 vm06 ceph-mon[49228]: osdmap e176: 8 total, 8 up, 8 in 2026-03-10T11:35:29.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:29 vm06 ceph-mon[57405]: osdmap e176: 8 total, 8 up, 8 in 2026-03-10T11:35:29.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:29 vm09 ceph-mon[54793]: osdmap e176: 8 total, 8 up, 8 in 2026-03-10T11:35:30.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:30 vm09 ceph-mon[54793]: pgmap v214: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 349 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:30.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:30 vm09 ceph-mon[54793]: osdmap e177: 8 total, 8 up, 8 in 2026-03-10T11:35:30.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:30 vm06 ceph-mon[49228]: pgmap v214: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 349 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:30.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:30 vm06 ceph-mon[49228]: osdmap e177: 8 total, 8 up, 8 in 2026-03-10T11:35:30.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:30 vm06 ceph-mon[57405]: pgmap v214: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 349 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:30.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:30 vm06 ceph-mon[57405]: osdmap e177: 8 total, 8 up, 8 in 2026-03-10T11:35:31.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:35:30 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:35:30] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:35:31.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:31 vm09 ceph-mon[54793]: osdmap e178: 8 total, 8 up, 8 in 2026-03-10T11:35:31.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:31 vm06 ceph-mon[49228]: osdmap e178: 8 total, 8 up, 8 in 2026-03-10T11:35:31.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:31 vm06 ceph-mon[57405]: osdmap e178: 8 total, 8 up, 8 in 2026-03-10T11:35:32.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:32 vm09 ceph-mon[54793]: pgmap v217: 196 pgs: 21 unknown, 175 active+clean; 455 KiB data, 349 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:32 vm09 ceph-mon[54793]: osdmap e179: 8 total, 8 up, 8 in 2026-03-10T11:35:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:32 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1015332252' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:32.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:32 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:32.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:32 vm06 ceph-mon[49228]: pgmap v217: 196 pgs: 21 unknown, 175 active+clean; 455 KiB data, 349 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:32.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:32 vm06 ceph-mon[49228]: osdmap e179: 8 total, 8 up, 8 in 2026-03-10T11:35:32.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:32 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1015332252' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:32.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:32 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:32 vm06 ceph-mon[57405]: pgmap v217: 196 pgs: 21 unknown, 175 active+clean; 455 KiB data, 349 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:32 vm06 ceph-mon[57405]: osdmap e179: 8 total, 8 up, 8 in 2026-03-10T11:35:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:32 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1015332252' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:32.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:32 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:33.348 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_snap_rollback PASSED [ 48%] 2026-03-10T11:35:33.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:33 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:33.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:33 vm09 ceph-mon[54793]: osdmap e180: 8 total, 8 up, 8 in 2026-03-10T11:35:33.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:33 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:33.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:33 vm06 ceph-mon[49228]: osdmap e180: 8 total, 8 up, 8 in 2026-03-10T11:35:33.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:33 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:33.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:33 vm06 ceph-mon[57405]: osdmap e180: 8 total, 8 up, 8 in 2026-03-10T11:35:34.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:34 vm09 ceph-mon[54793]: pgmap v220: 196 pgs: 196 active+clean; 455 KiB data, 350 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:35:34.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:34 vm09 ceph-mon[54793]: osdmap e181: 8 total, 8 up, 8 in 2026-03-10T11:35:34.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:34 vm06 ceph-mon[49228]: pgmap v220: 196 pgs: 196 active+clean; 455 KiB data, 350 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:35:34.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:34 vm06 ceph-mon[49228]: osdmap e181: 8 total, 8 up, 8 in 2026-03-10T11:35:34.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:34 vm06 ceph-mon[57405]: pgmap v220: 196 pgs: 196 active+clean; 455 KiB data, 350 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:35:34.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:34 vm06 ceph-mon[57405]: osdmap e181: 8 total, 8 up, 8 in 2026-03-10T11:35:35.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:35 vm09 ceph-mon[54793]: osdmap e182: 8 total, 8 up, 8 in 2026-03-10T11:35:35.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:35 vm06 ceph-mon[49228]: osdmap e182: 8 total, 8 up, 8 in 2026-03-10T11:35:35.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:35 vm06 ceph-mon[57405]: osdmap e182: 8 total, 8 up, 8 in 2026-03-10T11:35:36.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:36 vm09 ceph-mon[54793]: pgmap v223: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 350 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:36.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:36 vm09 ceph-mon[54793]: osdmap e183: 8 total, 8 up, 8 in 2026-03-10T11:35:36.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:36 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:36.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:35:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:35:36.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:36 vm06 ceph-mon[49228]: pgmap v223: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 350 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:36.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:36 vm06 ceph-mon[49228]: osdmap e183: 8 total, 8 up, 8 in 2026-03-10T11:35:36.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:36 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:36.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:36 vm06 ceph-mon[57405]: pgmap v223: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 350 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:36.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:36 vm06 ceph-mon[57405]: osdmap e183: 8 total, 8 up, 8 in 2026-03-10T11:35:36.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:36 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:37.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:37 vm09 ceph-mon[54793]: osdmap e184: 8 total, 8 up, 8 in 2026-03-10T11:35:37.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:37 vm09 ceph-mon[54793]: pgmap v226: 196 pgs: 196 active+clean; 455 KiB data, 385 MiB used, 160 GiB / 160 GiB avail; 1023 B/s rd, 0 op/s 2026-03-10T11:35:37.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:37 vm06 ceph-mon[49228]: osdmap e184: 8 total, 8 up, 8 in 2026-03-10T11:35:37.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:37 vm06 ceph-mon[49228]: pgmap v226: 196 pgs: 196 active+clean; 455 KiB data, 385 MiB used, 160 GiB / 160 GiB avail; 1023 B/s rd, 0 op/s 2026-03-10T11:35:37.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:37 vm06 ceph-mon[57405]: osdmap e184: 8 total, 8 up, 8 in 2026-03-10T11:35:37.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:37 vm06 ceph-mon[57405]: pgmap v226: 196 pgs: 196 active+clean; 455 KiB data, 385 MiB used, 160 GiB / 160 GiB avail; 1023 B/s rd, 0 op/s 2026-03-10T11:35:38.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:38 vm09 ceph-mon[54793]: osdmap e185: 8 total, 8 up, 8 in 2026-03-10T11:35:38.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:38 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1184908691' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:38.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:38 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:38.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:38 vm06 ceph-mon[49228]: osdmap e185: 8 total, 8 up, 8 in 2026-03-10T11:35:38.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:38 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1184908691' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:38.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:38 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:38.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:38 vm06 ceph-mon[57405]: osdmap e185: 8 total, 8 up, 8 in 2026-03-10T11:35:38.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:38 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1184908691' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:38.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:38 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:39.444 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_snap_rollback_removed PASSED [ 49%] 2026-03-10T11:35:39.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:39 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:39.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:39 vm09 ceph-mon[54793]: osdmap e186: 8 total, 8 up, 8 in 2026-03-10T11:35:39.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:39 vm09 ceph-mon[54793]: pgmap v229: 196 pgs: 196 active+clean; 455 KiB data, 385 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:35:39.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:39 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:39.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:39 vm06 ceph-mon[49228]: osdmap e186: 8 total, 8 up, 8 in 2026-03-10T11:35:39.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:39 vm06 ceph-mon[49228]: pgmap v229: 196 pgs: 196 active+clean; 455 KiB data, 385 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:35:39.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:39 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:39.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:39 vm06 ceph-mon[57405]: osdmap e186: 8 total, 8 up, 8 in 2026-03-10T11:35:39.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:39 vm06 ceph-mon[57405]: pgmap v229: 196 pgs: 196 active+clean; 455 KiB data, 385 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:35:40.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:40 vm09 ceph-mon[54793]: osdmap e187: 8 total, 8 up, 8 in 2026-03-10T11:35:40.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:40 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:40.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:40 vm06 ceph-mon[49228]: osdmap e187: 8 total, 8 up, 8 in 2026-03-10T11:35:40.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:40 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:40.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:40 vm06 ceph-mon[57405]: osdmap e187: 8 total, 8 up, 8 in 2026-03-10T11:35:40.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:40 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:41.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:35:40 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:35:40] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:35:41.742 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:41 vm09 ceph-mon[54793]: osdmap e188: 8 total, 8 up, 8 in 2026-03-10T11:35:41.742 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:41 vm09 ceph-mon[54793]: pgmap v232: 196 pgs: 19 creating+peering, 13 unknown, 164 active+clean; 455 KiB data, 390 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:41.742 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:41 vm09 ceph-mon[54793]: osdmap e189: 8 total, 8 up, 8 in 2026-03-10T11:35:41.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:41 vm06 ceph-mon[49228]: osdmap e188: 8 total, 8 up, 8 in 2026-03-10T11:35:41.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:41 vm06 ceph-mon[49228]: pgmap v232: 196 pgs: 19 creating+peering, 13 unknown, 164 active+clean; 455 KiB data, 390 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:41.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:41 vm06 ceph-mon[49228]: osdmap e189: 8 total, 8 up, 8 in 2026-03-10T11:35:41.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:41 vm06 ceph-mon[57405]: osdmap e188: 8 total, 8 up, 8 in 2026-03-10T11:35:41.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:41 vm06 ceph-mon[57405]: pgmap v232: 196 pgs: 19 creating+peering, 13 unknown, 164 active+clean; 455 KiB data, 390 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:41.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:41 vm06 ceph-mon[57405]: osdmap e189: 8 total, 8 up, 8 in 2026-03-10T11:35:43.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:43 vm09 ceph-mon[54793]: osdmap e190: 8 total, 8 up, 8 in 2026-03-10T11:35:43.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:43 vm09 ceph-mon[54793]: pgmap v235: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 408 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:43.730 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:43 vm06 ceph-mon[49228]: osdmap e190: 8 total, 8 up, 8 in 2026-03-10T11:35:43.730 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:43 vm06 ceph-mon[49228]: pgmap v235: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 408 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:43.730 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:43 vm06 ceph-mon[57405]: osdmap e190: 8 total, 8 up, 8 in 2026-03-10T11:35:43.730 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:43 vm06 ceph-mon[57405]: pgmap v235: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 408 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:44.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:44 vm06 ceph-mon[49228]: osdmap e191: 8 total, 8 up, 8 in 2026-03-10T11:35:44.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:44 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3041962692' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:44.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:44 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:44.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:44 vm06 ceph-mon[57405]: osdmap e191: 8 total, 8 up, 8 in 2026-03-10T11:35:44.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:44 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3041962692' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:44.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:44 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:44.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:44 vm09 ceph-mon[54793]: osdmap e191: 8 total, 8 up, 8 in 2026-03-10T11:35:44.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:44 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3041962692' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:44.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:44 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:45.506 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_snap_read PASSED [ 50%] 2026-03-10T11:35:45.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:45 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:45.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:45 vm06 ceph-mon[49228]: osdmap e192: 8 total, 8 up, 8 in 2026-03-10T11:35:45.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:45 vm06 ceph-mon[49228]: pgmap v238: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 408 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:45.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:45 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:45.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:45 vm06 ceph-mon[57405]: osdmap e192: 8 total, 8 up, 8 in 2026-03-10T11:35:45.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:45 vm06 ceph-mon[57405]: pgmap v238: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 408 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:45.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:45 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:45.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:45 vm09 ceph-mon[54793]: osdmap e192: 8 total, 8 up, 8 in 2026-03-10T11:35:45.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:45 vm09 ceph-mon[54793]: pgmap v238: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 408 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:46.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:35:46 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:35:46.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:46 vm06 ceph-mon[49228]: osdmap e193: 8 total, 8 up, 8 in 2026-03-10T11:35:46.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:46 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:46.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:46 vm06 ceph-mon[57405]: osdmap e193: 8 total, 8 up, 8 in 2026-03-10T11:35:46.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:46 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:46.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:46 vm09 ceph-mon[54793]: osdmap e193: 8 total, 8 up, 8 in 2026-03-10T11:35:46.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:46 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:47.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:47 vm06 ceph-mon[49228]: osdmap e194: 8 total, 8 up, 8 in 2026-03-10T11:35:47.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:47 vm06 ceph-mon[49228]: pgmap v241: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 444 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:47.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:47 vm06 ceph-mon[57405]: osdmap e194: 8 total, 8 up, 8 in 2026-03-10T11:35:47.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:47 vm06 ceph-mon[57405]: pgmap v241: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 444 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:47.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:47 vm09 ceph-mon[54793]: osdmap e194: 8 total, 8 up, 8 in 2026-03-10T11:35:47.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:47 vm09 ceph-mon[54793]: pgmap v241: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 444 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:48.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:48 vm06 ceph-mon[49228]: osdmap e195: 8 total, 8 up, 8 in 2026-03-10T11:35:48.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:48 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1013990807' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:48.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:48 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:48.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:48 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:48.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:48 vm06 ceph-mon[49228]: osdmap e196: 8 total, 8 up, 8 in 2026-03-10T11:35:48.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:48 vm06 ceph-mon[57405]: osdmap e195: 8 total, 8 up, 8 in 2026-03-10T11:35:48.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:48 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1013990807' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:48.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:48 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:48.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:48 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:48.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:48 vm06 ceph-mon[57405]: osdmap e196: 8 total, 8 up, 8 in 2026-03-10T11:35:48.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:48 vm09 ceph-mon[54793]: osdmap e195: 8 total, 8 up, 8 in 2026-03-10T11:35:48.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:48 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1013990807' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:48.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:48 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:48.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:48 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:48.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:48 vm09 ceph-mon[54793]: osdmap e196: 8 total, 8 up, 8 in 2026-03-10T11:35:49.511 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_set_omap PASSED [ 51%] 2026-03-10T11:35:49.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:49 vm09 ceph-mon[54793]: pgmap v244: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 444 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:49.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:49 vm09 ceph-mon[54793]: osdmap e197: 8 total, 8 up, 8 in 2026-03-10T11:35:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:49 vm06 ceph-mon[49228]: pgmap v244: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 444 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:50.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:49 vm06 ceph-mon[49228]: osdmap e197: 8 total, 8 up, 8 in 2026-03-10T11:35:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:49 vm06 ceph-mon[57405]: pgmap v244: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 444 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:50.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:49 vm06 ceph-mon[57405]: osdmap e197: 8 total, 8 up, 8 in 2026-03-10T11:35:51.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:35:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:35:50] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:35:51.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:51 vm09 ceph-mon[54793]: osdmap e198: 8 total, 8 up, 8 in 2026-03-10T11:35:51.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:51 vm09 ceph-mon[54793]: pgmap v247: 196 pgs: 20 creating+peering, 12 unknown, 164 active+clean; 455 KiB data, 448 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:52.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:51 vm06 ceph-mon[49228]: osdmap e198: 8 total, 8 up, 8 in 2026-03-10T11:35:52.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:51 vm06 ceph-mon[49228]: pgmap v247: 196 pgs: 20 creating+peering, 12 unknown, 164 active+clean; 455 KiB data, 448 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:52.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:51 vm06 ceph-mon[57405]: osdmap e198: 8 total, 8 up, 8 in 2026-03-10T11:35:52.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:51 vm06 ceph-mon[57405]: pgmap v247: 196 pgs: 20 creating+peering, 12 unknown, 164 active+clean; 455 KiB data, 448 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:52.920 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:52 vm09 ceph-mon[54793]: osdmap e199: 8 total, 8 up, 8 in 2026-03-10T11:35:52.920 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:52 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3717570321' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:52.920 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:52 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:35:52.920 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:52 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:35:52.920 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:52 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:35:52.920 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:52 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[49228]: osdmap e199: 8 total, 8 up, 8 in 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3717570321' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[57405]: osdmap e199: 8 total, 8 up, 8 in 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3717570321' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:35:53.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:52 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:35:53.579 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_set_omap_aio PASSED [ 52%] 2026-03-10T11:35:53.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:53 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3717570321' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:53.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:53 vm09 ceph-mon[54793]: osdmap e200: 8 total, 8 up, 8 in 2026-03-10T11:35:53.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:53 vm09 ceph-mon[54793]: pgmap v250: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 456 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:54.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:53 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3717570321' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:54.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:53 vm06 ceph-mon[49228]: osdmap e200: 8 total, 8 up, 8 in 2026-03-10T11:35:54.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:53 vm06 ceph-mon[49228]: pgmap v250: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 456 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:54.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:53 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3717570321' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:54.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:53 vm06 ceph-mon[57405]: osdmap e200: 8 total, 8 up, 8 in 2026-03-10T11:35:54.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:53 vm06 ceph-mon[57405]: pgmap v250: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 456 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:54.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:54 vm09 ceph-mon[54793]: osdmap e201: 8 total, 8 up, 8 in 2026-03-10T11:35:55.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:54 vm06 ceph-mon[49228]: osdmap e201: 8 total, 8 up, 8 in 2026-03-10T11:35:55.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:54 vm06 ceph-mon[57405]: osdmap e201: 8 total, 8 up, 8 in 2026-03-10T11:35:55.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:55 vm09 ceph-mon[54793]: osdmap e202: 8 total, 8 up, 8 in 2026-03-10T11:35:55.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:55 vm09 ceph-mon[54793]: pgmap v253: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 456 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:55.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:55 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:56.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:55 vm06 ceph-mon[49228]: osdmap e202: 8 total, 8 up, 8 in 2026-03-10T11:35:56.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:55 vm06 ceph-mon[49228]: pgmap v253: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 456 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:56.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:55 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:56.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:55 vm06 ceph-mon[57405]: osdmap e202: 8 total, 8 up, 8 in 2026-03-10T11:35:56.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:55 vm06 ceph-mon[57405]: pgmap v253: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 456 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:35:56.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:55 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:35:56.479 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:35:56 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:35:56.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:56 vm09 ceph-mon[54793]: osdmap e203: 8 total, 8 up, 8 in 2026-03-10T11:35:56.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:56 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2503320135' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:56.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:56 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:57.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:56 vm06 ceph-mon[49228]: osdmap e203: 8 total, 8 up, 8 in 2026-03-10T11:35:57.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:56 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2503320135' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:57.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:56 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:57.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:56 vm06 ceph-mon[57405]: osdmap e203: 8 total, 8 up, 8 in 2026-03-10T11:35:57.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:56 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2503320135' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:35:57.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:56 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:35:57.676 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_write_ops PASSED [ 53%] 2026-03-10T11:35:57.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:57 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2503320135' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:57.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:57 vm09 ceph-mon[54793]: osdmap e204: 8 total, 8 up, 8 in 2026-03-10T11:35:57.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:57 vm09 ceph-mon[54793]: pgmap v256: 196 pgs: 196 active+clean; 455 KiB data, 457 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 1.5 KiB/s wr, 4 op/s 2026-03-10T11:35:58.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:57 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2503320135' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:58.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:57 vm06 ceph-mon[49228]: osdmap e204: 8 total, 8 up, 8 in 2026-03-10T11:35:58.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:57 vm06 ceph-mon[49228]: pgmap v256: 196 pgs: 196 active+clean; 455 KiB data, 457 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 1.5 KiB/s wr, 4 op/s 2026-03-10T11:35:58.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:57 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2503320135' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:35:58.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:57 vm06 ceph-mon[57405]: osdmap e204: 8 total, 8 up, 8 in 2026-03-10T11:35:58.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:57 vm06 ceph-mon[57405]: pgmap v256: 196 pgs: 196 active+clean; 455 KiB data, 457 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 1.5 KiB/s wr, 4 op/s 2026-03-10T11:35:58.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:58 vm09 ceph-mon[54793]: osdmap e205: 8 total, 8 up, 8 in 2026-03-10T11:35:59.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:58 vm06 ceph-mon[49228]: osdmap e205: 8 total, 8 up, 8 in 2026-03-10T11:35:59.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:58 vm06 ceph-mon[57405]: osdmap e205: 8 total, 8 up, 8 in 2026-03-10T11:35:59.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:59 vm09 ceph-mon[54793]: osdmap e206: 8 total, 8 up, 8 in 2026-03-10T11:35:59.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:59 vm09 ceph-mon[54793]: pgmap v259: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 457 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:35:59.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:35:59 vm09 ceph-mon[54793]: osdmap e207: 8 total, 8 up, 8 in 2026-03-10T11:36:00.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:59 vm06 ceph-mon[49228]: osdmap e206: 8 total, 8 up, 8 in 2026-03-10T11:36:00.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:59 vm06 ceph-mon[49228]: pgmap v259: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 457 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:00.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:35:59 vm06 ceph-mon[49228]: osdmap e207: 8 total, 8 up, 8 in 2026-03-10T11:36:00.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:59 vm06 ceph-mon[57405]: osdmap e206: 8 total, 8 up, 8 in 2026-03-10T11:36:00.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:59 vm06 ceph-mon[57405]: pgmap v259: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 457 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:00.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:35:59 vm06 ceph-mon[57405]: osdmap e207: 8 total, 8 up, 8 in 2026-03-10T11:36:00.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:00 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/5344104' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:01.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:36:00 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:36:00] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:36:01.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:00 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/5344104' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:01.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:00 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/5344104' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:01.713 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_execute_op PASSED [ 54%] 2026-03-10T11:36:02.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:01 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/5344104' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:02.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:01 vm06 ceph-mon[49228]: osdmap e208: 8 total, 8 up, 8 in 2026-03-10T11:36:02.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:01 vm06 ceph-mon[49228]: pgmap v262: 196 pgs: 11 unknown, 185 active+clean; 455 KiB data, 461 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 0 op/s 2026-03-10T11:36:02.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:01 vm06 ceph-mon[49228]: osdmap e209: 8 total, 8 up, 8 in 2026-03-10T11:36:02.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:01 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/5344104' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:02.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:01 vm06 ceph-mon[57405]: osdmap e208: 8 total, 8 up, 8 in 2026-03-10T11:36:02.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:01 vm06 ceph-mon[57405]: pgmap v262: 196 pgs: 11 unknown, 185 active+clean; 455 KiB data, 461 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 0 op/s 2026-03-10T11:36:02.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:01 vm06 ceph-mon[57405]: osdmap e209: 8 total, 8 up, 8 in 2026-03-10T11:36:02.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:01 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/5344104' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:02.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:01 vm09 ceph-mon[54793]: osdmap e208: 8 total, 8 up, 8 in 2026-03-10T11:36:02.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:01 vm09 ceph-mon[54793]: pgmap v262: 196 pgs: 11 unknown, 185 active+clean; 455 KiB data, 461 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 0 op/s 2026-03-10T11:36:02.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:01 vm09 ceph-mon[54793]: osdmap e209: 8 total, 8 up, 8 in 2026-03-10T11:36:04.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:03 vm06 ceph-mon[49228]: osdmap e210: 8 total, 8 up, 8 in 2026-03-10T11:36:04.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:03 vm06 ceph-mon[49228]: pgmap v265: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:04.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:03 vm06 ceph-mon[57405]: osdmap e210: 8 total, 8 up, 8 in 2026-03-10T11:36:04.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:03 vm06 ceph-mon[57405]: pgmap v265: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:04.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:03 vm09 ceph-mon[54793]: osdmap e210: 8 total, 8 up, 8 in 2026-03-10T11:36:04.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:03 vm09 ceph-mon[54793]: pgmap v265: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:05.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:04 vm06 ceph-mon[49228]: osdmap e211: 8 total, 8 up, 8 in 2026-03-10T11:36:05.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:04 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1221266996' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:05.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:04 vm06 ceph-mon[57405]: osdmap e211: 8 total, 8 up, 8 in 2026-03-10T11:36:05.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:04 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1221266996' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:04 vm09 ceph-mon[54793]: osdmap e211: 8 total, 8 up, 8 in 2026-03-10T11:36:05.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:04 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1221266996' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:05.806 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_writesame_op PASSED [ 56%] 2026-03-10T11:36:06.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:05 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1221266996' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:06.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:05 vm09 ceph-mon[54793]: osdmap e212: 8 total, 8 up, 8 in 2026-03-10T11:36:06.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:05 vm09 ceph-mon[54793]: pgmap v268: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:06.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:05 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1221266996' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:06.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:05 vm06 ceph-mon[49228]: osdmap e212: 8 total, 8 up, 8 in 2026-03-10T11:36:06.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:05 vm06 ceph-mon[49228]: pgmap v268: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:06.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:05 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1221266996' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:06.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:05 vm06 ceph-mon[57405]: osdmap e212: 8 total, 8 up, 8 in 2026-03-10T11:36:06.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:05 vm06 ceph-mon[57405]: pgmap v268: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:06.730 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:36:06 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:36:07.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:06 vm09 ceph-mon[54793]: osdmap e213: 8 total, 8 up, 8 in 2026-03-10T11:36:07.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:06 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:07.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:06 vm06 ceph-mon[49228]: osdmap e213: 8 total, 8 up, 8 in 2026-03-10T11:36:07.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:06 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:07.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:06 vm06 ceph-mon[57405]: osdmap e213: 8 total, 8 up, 8 in 2026-03-10T11:36:07.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:06 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:08.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:07 vm09 ceph-mon[54793]: osdmap e214: 8 total, 8 up, 8 in 2026-03-10T11:36:08.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:07 vm09 ceph-mon[54793]: pgmap v271: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:08.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:07 vm09 ceph-mon[54793]: osdmap e215: 8 total, 8 up, 8 in 2026-03-10T11:36:08.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:07 vm06 ceph-mon[49228]: osdmap e214: 8 total, 8 up, 8 in 2026-03-10T11:36:08.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:07 vm06 ceph-mon[49228]: pgmap v271: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:08.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:07 vm06 ceph-mon[49228]: osdmap e215: 8 total, 8 up, 8 in 2026-03-10T11:36:08.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:07 vm06 ceph-mon[57405]: osdmap e214: 8 total, 8 up, 8 in 2026-03-10T11:36:08.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:07 vm06 ceph-mon[57405]: pgmap v271: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:08.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:07 vm06 ceph-mon[57405]: osdmap e215: 8 total, 8 up, 8 in 2026-03-10T11:36:09.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:08 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/587341621' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:09.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:08 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/587341621' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:09.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:08 vm09 ceph-mon[54793]: osdmap e216: 8 total, 8 up, 8 in 2026-03-10T11:36:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:08 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/587341621' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:08 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/587341621' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:09.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:08 vm06 ceph-mon[49228]: osdmap e216: 8 total, 8 up, 8 in 2026-03-10T11:36:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:08 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/587341621' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:08 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/587341621' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:09.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:08 vm06 ceph-mon[57405]: osdmap e216: 8 total, 8 up, 8 in 2026-03-10T11:36:09.826 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_get_omap_vals_by_keys PASSED [ 57%] 2026-03-10T11:36:10.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:09 vm09 ceph-mon[54793]: pgmap v274: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:10.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:09 vm09 ceph-mon[54793]: osdmap e217: 8 total, 8 up, 8 in 2026-03-10T11:36:10.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:09 vm06 ceph-mon[49228]: pgmap v274: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:10.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:09 vm06 ceph-mon[49228]: osdmap e217: 8 total, 8 up, 8 in 2026-03-10T11:36:10.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:09 vm06 ceph-mon[57405]: pgmap v274: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 462 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:10.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:09 vm06 ceph-mon[57405]: osdmap e217: 8 total, 8 up, 8 in 2026-03-10T11:36:11.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:10 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:11.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:10 vm06 ceph-mon[49228]: osdmap e218: 8 total, 8 up, 8 in 2026-03-10T11:36:11.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:36:10 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:36:10] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:36:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:10 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:10 vm06 ceph-mon[57405]: osdmap e218: 8 total, 8 up, 8 in 2026-03-10T11:36:11.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:10 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:11.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:10 vm09 ceph-mon[54793]: osdmap e218: 8 total, 8 up, 8 in 2026-03-10T11:36:12.131 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:11 vm09 ceph-mon[54793]: pgmap v277: 196 pgs: 16 creating+peering, 16 unknown, 164 active+clean; 455 KiB data, 463 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:12.131 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:11 vm09 ceph-mon[54793]: osdmap e219: 8 total, 8 up, 8 in 2026-03-10T11:36:12.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:11 vm06 ceph-mon[49228]: pgmap v277: 196 pgs: 16 creating+peering, 16 unknown, 164 active+clean; 455 KiB data, 463 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:12.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:11 vm06 ceph-mon[49228]: osdmap e219: 8 total, 8 up, 8 in 2026-03-10T11:36:12.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:11 vm06 ceph-mon[57405]: pgmap v277: 196 pgs: 16 creating+peering, 16 unknown, 164 active+clean; 455 KiB data, 463 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:12.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:11 vm06 ceph-mon[57405]: osdmap e219: 8 total, 8 up, 8 in 2026-03-10T11:36:13.182 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:12 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2932073287' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:13.182 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:12 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:13.182 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:12 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:13.182 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:12 vm06 ceph-mon[49228]: osdmap e220: 8 total, 8 up, 8 in 2026-03-10T11:36:13.182 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:12 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2932073287' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:13.182 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:12 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:13.182 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:12 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:13.182 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:12 vm06 ceph-mon[57405]: osdmap e220: 8 total, 8 up, 8 in 2026-03-10T11:36:13.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:12 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2932073287' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:13.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:12 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:13.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:12 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:13.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:12 vm09 ceph-mon[54793]: osdmap e220: 8 total, 8 up, 8 in 2026-03-10T11:36:13.856 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_get_omap_keys PASSED [ 58%] 2026-03-10T11:36:14.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:13 vm09 ceph-mon[54793]: pgmap v280: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 463 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:14.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:13 vm09 ceph-mon[54793]: osdmap e221: 8 total, 8 up, 8 in 2026-03-10T11:36:14.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:13 vm06 ceph-mon[49228]: pgmap v280: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 463 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:14.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:13 vm06 ceph-mon[49228]: osdmap e221: 8 total, 8 up, 8 in 2026-03-10T11:36:14.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:13 vm06 ceph-mon[57405]: pgmap v280: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 463 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:14.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:13 vm06 ceph-mon[57405]: osdmap e221: 8 total, 8 up, 8 in 2026-03-10T11:36:16.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:15 vm09 ceph-mon[54793]: osdmap e222: 8 total, 8 up, 8 in 2026-03-10T11:36:16.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:15 vm09 ceph-mon[54793]: pgmap v283: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 463 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:16.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:15 vm06 ceph-mon[49228]: osdmap e222: 8 total, 8 up, 8 in 2026-03-10T11:36:16.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:15 vm06 ceph-mon[49228]: pgmap v283: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 463 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:16.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:15 vm06 ceph-mon[57405]: osdmap e222: 8 total, 8 up, 8 in 2026-03-10T11:36:16.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:15 vm06 ceph-mon[57405]: pgmap v283: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 463 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:16.730 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:36:16 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:36:17.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:16 vm09 ceph-mon[54793]: osdmap e223: 8 total, 8 up, 8 in 2026-03-10T11:36:17.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:16 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/4197430367' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:17.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:16 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:17.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:16 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:17.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:16 vm06 ceph-mon[49228]: osdmap e223: 8 total, 8 up, 8 in 2026-03-10T11:36:17.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:16 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/4197430367' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:17.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:16 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:17.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:16 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:17.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:16 vm06 ceph-mon[57405]: osdmap e223: 8 total, 8 up, 8 in 2026-03-10T11:36:17.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:16 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/4197430367' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:17.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:16 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:17.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:16 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:17.939 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_clear_omap PASSED [ 59%] 2026-03-10T11:36:18.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:17 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:18.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:17 vm09 ceph-mon[54793]: osdmap e224: 8 total, 8 up, 8 in 2026-03-10T11:36:18.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:17 vm09 ceph-mon[54793]: pgmap v286: 196 pgs: 196 active+clean; 455 KiB data, 464 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:36:18.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:17 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:17 vm06 ceph-mon[49228]: osdmap e224: 8 total, 8 up, 8 in 2026-03-10T11:36:18.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:17 vm06 ceph-mon[49228]: pgmap v286: 196 pgs: 196 active+clean; 455 KiB data, 464 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:36:18.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:17 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:18.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:17 vm06 ceph-mon[57405]: osdmap e224: 8 total, 8 up, 8 in 2026-03-10T11:36:18.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:17 vm06 ceph-mon[57405]: pgmap v286: 196 pgs: 196 active+clean; 455 KiB data, 464 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:36:19.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:18 vm09 ceph-mon[54793]: osdmap e225: 8 total, 8 up, 8 in 2026-03-10T11:36:19.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:18 vm06 ceph-mon[49228]: osdmap e225: 8 total, 8 up, 8 in 2026-03-10T11:36:19.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:18 vm06 ceph-mon[57405]: osdmap e225: 8 total, 8 up, 8 in 2026-03-10T11:36:20.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:19 vm09 ceph-mon[54793]: osdmap e226: 8 total, 8 up, 8 in 2026-03-10T11:36:20.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:19 vm09 ceph-mon[54793]: pgmap v289: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 464 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:20.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:19 vm06 ceph-mon[49228]: osdmap e226: 8 total, 8 up, 8 in 2026-03-10T11:36:20.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:19 vm06 ceph-mon[49228]: pgmap v289: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 464 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:20.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:19 vm06 ceph-mon[57405]: osdmap e226: 8 total, 8 up, 8 in 2026-03-10T11:36:20.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:19 vm06 ceph-mon[57405]: pgmap v289: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 464 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:21.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:20 vm06 ceph-mon[49228]: osdmap e227: 8 total, 8 up, 8 in 2026-03-10T11:36:21.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:20 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2576180273' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:21.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:20 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:21.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:36:20 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:36:20] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:36:21.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:20 vm06 ceph-mon[57405]: osdmap e227: 8 total, 8 up, 8 in 2026-03-10T11:36:21.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:20 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2576180273' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:21.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:20 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:21.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:20 vm09 ceph-mon[54793]: osdmap e227: 8 total, 8 up, 8 in 2026-03-10T11:36:21.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:20 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2576180273' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:21.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:20 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:21.999 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_remove_omap_range2 PASSED [ 60%] 2026-03-10T11:36:22.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:21 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:22.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:21 vm06 ceph-mon[49228]: osdmap e228: 8 total, 8 up, 8 in 2026-03-10T11:36:22.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:21 vm06 ceph-mon[49228]: pgmap v292: 196 pgs: 12 unknown, 184 active+clean; 455 KiB data, 464 MiB used, 160 GiB / 160 GiB avail; 1023 B/s rd, 0 op/s 2026-03-10T11:36:22.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:21 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:22.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:21 vm06 ceph-mon[57405]: osdmap e228: 8 total, 8 up, 8 in 2026-03-10T11:36:22.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:21 vm06 ceph-mon[57405]: pgmap v292: 196 pgs: 12 unknown, 184 active+clean; 455 KiB data, 464 MiB used, 160 GiB / 160 GiB avail; 1023 B/s rd, 0 op/s 2026-03-10T11:36:22.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:21 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:22.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:21 vm09 ceph-mon[54793]: osdmap e228: 8 total, 8 up, 8 in 2026-03-10T11:36:22.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:21 vm09 ceph-mon[54793]: pgmap v292: 196 pgs: 12 unknown, 184 active+clean; 455 KiB data, 464 MiB used, 160 GiB / 160 GiB avail; 1023 B/s rd, 0 op/s 2026-03-10T11:36:23.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:22 vm06 ceph-mon[49228]: osdmap e229: 8 total, 8 up, 8 in 2026-03-10T11:36:23.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:22 vm06 ceph-mon[57405]: osdmap e229: 8 total, 8 up, 8 in 2026-03-10T11:36:23.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:22 vm09 ceph-mon[54793]: osdmap e229: 8 total, 8 up, 8 in 2026-03-10T11:36:24.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:24 vm06 ceph-mon[49228]: osdmap e230: 8 total, 8 up, 8 in 2026-03-10T11:36:24.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:24 vm06 ceph-mon[49228]: pgmap v295: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 468 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:24 vm06 ceph-mon[57405]: osdmap e230: 8 total, 8 up, 8 in 2026-03-10T11:36:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:24 vm06 ceph-mon[57405]: pgmap v295: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 468 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:24.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:24 vm09 ceph-mon[54793]: osdmap e230: 8 total, 8 up, 8 in 2026-03-10T11:36:24.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:24 vm09 ceph-mon[54793]: pgmap v295: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 468 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:25.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:25 vm06 ceph-mon[49228]: osdmap e231: 8 total, 8 up, 8 in 2026-03-10T11:36:25.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:25 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1190522086' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:25.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:25 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:25.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:25 vm06 ceph-mon[57405]: osdmap e231: 8 total, 8 up, 8 in 2026-03-10T11:36:25.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:25 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1190522086' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:25.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:25 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:25 vm09 ceph-mon[54793]: osdmap e231: 8 total, 8 up, 8 in 2026-03-10T11:36:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:25 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1190522086' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:25.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:25 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:26.032 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_omap_cmp PASSED [ 61%] 2026-03-10T11:36:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:26 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:26 vm09 ceph-mon[54793]: osdmap e232: 8 total, 8 up, 8 in 2026-03-10T11:36:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:26 vm09 ceph-mon[54793]: pgmap v298: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 468 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:26.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:26 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:26.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:36:26 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:36:26.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:26 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:26.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:26 vm06 ceph-mon[49228]: osdmap e232: 8 total, 8 up, 8 in 2026-03-10T11:36:26.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:26 vm06 ceph-mon[49228]: pgmap v298: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 468 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:26.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:26 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:26 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:26 vm06 ceph-mon[57405]: osdmap e232: 8 total, 8 up, 8 in 2026-03-10T11:36:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:26 vm06 ceph-mon[57405]: pgmap v298: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 468 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:26.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:26 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:27 vm09 ceph-mon[54793]: osdmap e233: 8 total, 8 up, 8 in 2026-03-10T11:36:27.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:27 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:27.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:27 vm06 ceph-mon[49228]: osdmap e233: 8 total, 8 up, 8 in 2026-03-10T11:36:27.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:27 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:27.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:27 vm06 ceph-mon[57405]: osdmap e233: 8 total, 8 up, 8 in 2026-03-10T11:36:27.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:27 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:28.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:28 vm09 ceph-mon[54793]: pgmap v300: 164 pgs: 164 active+clean; 455 KiB data, 477 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:28.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:28 vm09 ceph-mon[54793]: osdmap e234: 8 total, 8 up, 8 in 2026-03-10T11:36:28.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:28 vm06 ceph-mon[49228]: pgmap v300: 164 pgs: 164 active+clean; 455 KiB data, 477 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:28.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:28 vm06 ceph-mon[49228]: osdmap e234: 8 total, 8 up, 8 in 2026-03-10T11:36:28.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:28 vm06 ceph-mon[57405]: pgmap v300: 164 pgs: 164 active+clean; 455 KiB data, 477 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:28.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:28 vm06 ceph-mon[57405]: osdmap e234: 8 total, 8 up, 8 in 2026-03-10T11:36:29.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:29 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:29.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:29 vm09 ceph-mon[54793]: osdmap e235: 8 total, 8 up, 8 in 2026-03-10T11:36:29.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:29 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1820271033' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:29.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:29 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:29.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:29 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:29.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:29 vm06 ceph-mon[49228]: osdmap e235: 8 total, 8 up, 8 in 2026-03-10T11:36:29.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:29 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1820271033' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:29.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:29 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:29.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:29 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:29.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:29 vm06 ceph-mon[57405]: osdmap e235: 8 total, 8 up, 8 in 2026-03-10T11:36:29.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:29 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1820271033' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:29.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:29 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:30.080 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_cmpext_op PASSED [ 62%] 2026-03-10T11:36:30.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:30 vm09 ceph-mon[54793]: pgmap v303: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 477 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:30.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:30 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:30.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:30 vm09 ceph-mon[54793]: osdmap e236: 8 total, 8 up, 8 in 2026-03-10T11:36:30.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:30 vm06 ceph-mon[49228]: pgmap v303: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 477 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:30.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:30 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:30.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:30 vm06 ceph-mon[49228]: osdmap e236: 8 total, 8 up, 8 in 2026-03-10T11:36:30.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:30 vm06 ceph-mon[57405]: pgmap v303: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 477 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:30.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:30 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:30.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:30 vm06 ceph-mon[57405]: osdmap e236: 8 total, 8 up, 8 in 2026-03-10T11:36:31.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:36:30 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:36:30] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:36:31.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:31 vm09 ceph-mon[54793]: osdmap e237: 8 total, 8 up, 8 in 2026-03-10T11:36:31.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:31 vm06 ceph-mon[49228]: osdmap e237: 8 total, 8 up, 8 in 2026-03-10T11:36:31.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:31 vm06 ceph-mon[57405]: osdmap e237: 8 total, 8 up, 8 in 2026-03-10T11:36:32.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:32 vm09 ceph-mon[54793]: pgmap v306: 164 pgs: 164 active+clean; 455 KiB data, 477 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:32.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:32 vm09 ceph-mon[54793]: osdmap e238: 8 total, 8 up, 8 in 2026-03-10T11:36:32.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:32 vm06 ceph-mon[49228]: pgmap v306: 164 pgs: 164 active+clean; 455 KiB data, 477 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:32.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:32 vm06 ceph-mon[49228]: osdmap e238: 8 total, 8 up, 8 in 2026-03-10T11:36:32.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:32 vm06 ceph-mon[57405]: pgmap v306: 164 pgs: 164 active+clean; 455 KiB data, 477 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:32.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:32 vm06 ceph-mon[57405]: osdmap e238: 8 total, 8 up, 8 in 2026-03-10T11:36:33.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:33 vm09 ceph-mon[54793]: osdmap e239: 8 total, 8 up, 8 in 2026-03-10T11:36:33.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:33 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3784473389' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:33.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:33 vm06 ceph-mon[49228]: osdmap e239: 8 total, 8 up, 8 in 2026-03-10T11:36:33.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:33 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3784473389' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:33.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:33 vm06 ceph-mon[57405]: osdmap e239: 8 total, 8 up, 8 in 2026-03-10T11:36:33.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:33 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3784473389' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:34.148 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_xattrs_op PASSED [ 63%] 2026-03-10T11:36:34.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:34 vm09 ceph-mon[54793]: pgmap v309: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:34.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:34 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:34.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:34 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3784473389' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:34.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:34 vm09 ceph-mon[54793]: osdmap e240: 8 total, 8 up, 8 in 2026-03-10T11:36:34.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:34 vm06 ceph-mon[49228]: pgmap v309: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:34.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:34 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:34.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:34 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3784473389' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:34.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:34 vm06 ceph-mon[49228]: osdmap e240: 8 total, 8 up, 8 in 2026-03-10T11:36:34.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:34 vm06 ceph-mon[57405]: pgmap v309: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:34.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:34 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:34.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:34 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3784473389' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:34.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:34 vm06 ceph-mon[57405]: osdmap e240: 8 total, 8 up, 8 in 2026-03-10T11:36:35.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:35 vm09 ceph-mon[54793]: osdmap e241: 8 total, 8 up, 8 in 2026-03-10T11:36:35.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:35 vm06 ceph-mon[49228]: osdmap e241: 8 total, 8 up, 8 in 2026-03-10T11:36:35.530 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:35 vm06 ceph-mon[57405]: osdmap e241: 8 total, 8 up, 8 in 2026-03-10T11:36:36.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:36 vm09 ceph-mon[54793]: pgmap v312: 164 pgs: 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:36.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:36 vm09 ceph-mon[54793]: osdmap e242: 8 total, 8 up, 8 in 2026-03-10T11:36:36.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:36:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:36:36.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:36 vm06 ceph-mon[49228]: pgmap v312: 164 pgs: 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:36.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:36 vm06 ceph-mon[49228]: osdmap e242: 8 total, 8 up, 8 in 2026-03-10T11:36:36.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:36 vm06 ceph-mon[57405]: pgmap v312: 164 pgs: 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:36.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:36 vm06 ceph-mon[57405]: osdmap e242: 8 total, 8 up, 8 in 2026-03-10T11:36:37.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:37 vm09 ceph-mon[54793]: osdmap e243: 8 total, 8 up, 8 in 2026-03-10T11:36:37.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:37 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2809424689' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:37.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:37 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:37.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:37 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:37.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:37 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:37.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:37 vm09 ceph-mon[54793]: osdmap e244: 8 total, 8 up, 8 in 2026-03-10T11:36:37.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[49228]: osdmap e243: 8 total, 8 up, 8 in 2026-03-10T11:36:37.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2809424689' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:37.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:37.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:37.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:37.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[49228]: osdmap e244: 8 total, 8 up, 8 in 2026-03-10T11:36:37.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[57405]: osdmap e243: 8 total, 8 up, 8 in 2026-03-10T11:36:37.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2809424689' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:37.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:37.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:37.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:37.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:37 vm06 ceph-mon[57405]: osdmap e244: 8 total, 8 up, 8 in 2026-03-10T11:36:38.177 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_locator PASSED [ 64%] 2026-03-10T11:36:38.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:38 vm09 ceph-mon[54793]: pgmap v315: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:38.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:38 vm09 ceph-mon[54793]: osdmap e245: 8 total, 8 up, 8 in 2026-03-10T11:36:38.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:38 vm06 ceph-mon[49228]: pgmap v315: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:38.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:38 vm06 ceph-mon[49228]: osdmap e245: 8 total, 8 up, 8 in 2026-03-10T11:36:38.530 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:38 vm06 ceph-mon[57405]: pgmap v315: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:38.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:38 vm06 ceph-mon[57405]: osdmap e245: 8 total, 8 up, 8 in 2026-03-10T11:36:39.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:39 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:39.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:39 vm06 ceph-mon[49228]: osdmap e246: 8 total, 8 up, 8 in 2026-03-10T11:36:39.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:39 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:39.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:39 vm06 ceph-mon[57405]: osdmap e246: 8 total, 8 up, 8 in 2026-03-10T11:36:39.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:39 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:39.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:39 vm09 ceph-mon[54793]: osdmap e246: 8 total, 8 up, 8 in 2026-03-10T11:36:40.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:40 vm06 ceph-mon[49228]: pgmap v318: 164 pgs: 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:40.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:40 vm06 ceph-mon[49228]: osdmap e247: 8 total, 8 up, 8 in 2026-03-10T11:36:40.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:40 vm06 ceph-mon[57405]: pgmap v318: 164 pgs: 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:40.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:40 vm06 ceph-mon[57405]: osdmap e247: 8 total, 8 up, 8 in 2026-03-10T11:36:40.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:40 vm09 ceph-mon[54793]: pgmap v318: 164 pgs: 164 active+clean; 455 KiB data, 478 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:40.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:40 vm09 ceph-mon[54793]: osdmap e247: 8 total, 8 up, 8 in 2026-03-10T11:36:41.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:36:40 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:36:40] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:36:41.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:41 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1436882363' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:41.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:41 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:41.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:41 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:41.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:41 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:41.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:41 vm06 ceph-mon[49228]: osdmap e248: 8 total, 8 up, 8 in 2026-03-10T11:36:41.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:41 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1436882363' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:41.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:41 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:41.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:41 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:41.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:41 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:41.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:41 vm06 ceph-mon[57405]: osdmap e248: 8 total, 8 up, 8 in 2026-03-10T11:36:41.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:41 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1436882363' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:41.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:41 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:41.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:41 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:41.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:41 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:41.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:41 vm09 ceph-mon[54793]: osdmap e248: 8 total, 8 up, 8 in 2026-03-10T11:36:42.208 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_operate_aio_write_op PASSED [ 65%] 2026-03-10T11:36:42.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:42 vm06 ceph-mon[49228]: pgmap v321: 196 pgs: 7 unknown, 189 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:36:42.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:42 vm06 ceph-mon[49228]: osdmap e249: 8 total, 8 up, 8 in 2026-03-10T11:36:42.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:42 vm06 ceph-mon[57405]: pgmap v321: 196 pgs: 7 unknown, 189 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:36:42.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:42 vm06 ceph-mon[57405]: osdmap e249: 8 total, 8 up, 8 in 2026-03-10T11:36:42.631 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:42 vm09 ceph-mon[54793]: pgmap v321: 196 pgs: 7 unknown, 189 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:36:42.631 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:42 vm09 ceph-mon[54793]: osdmap e249: 8 total, 8 up, 8 in 2026-03-10T11:36:44.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:44 vm06 ceph-mon[49228]: pgmap v324: 164 pgs: 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:44.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:44 vm06 ceph-mon[49228]: osdmap e250: 8 total, 8 up, 8 in 2026-03-10T11:36:44.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:44 vm06 ceph-mon[57405]: pgmap v324: 164 pgs: 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:44.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:44 vm06 ceph-mon[57405]: osdmap e250: 8 total, 8 up, 8 in 2026-03-10T11:36:44.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:44 vm09 ceph-mon[54793]: pgmap v324: 164 pgs: 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:44.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:44 vm09 ceph-mon[54793]: osdmap e250: 8 total, 8 up, 8 in 2026-03-10T11:36:45.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:45 vm06 ceph-mon[49228]: osdmap e251: 8 total, 8 up, 8 in 2026-03-10T11:36:45.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:45 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/615563678' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:45.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:45 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:45.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:45 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:45.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:45 vm06 ceph-mon[49228]: osdmap e252: 8 total, 8 up, 8 in 2026-03-10T11:36:45.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:45 vm06 ceph-mon[57405]: osdmap e251: 8 total, 8 up, 8 in 2026-03-10T11:36:45.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:45 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/615563678' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:45.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:45 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:45.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:45 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:45.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:45 vm06 ceph-mon[57405]: osdmap e252: 8 total, 8 up, 8 in 2026-03-10T11:36:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:45 vm09 ceph-mon[54793]: osdmap e251: 8 total, 8 up, 8 in 2026-03-10T11:36:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:45 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/615563678' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:45 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:45 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:45.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:45 vm09 ceph-mon[54793]: osdmap e252: 8 total, 8 up, 8 in 2026-03-10T11:36:46.247 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_write PASSED [ 67%] 2026-03-10T11:36:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:46 vm06 ceph-mon[49228]: pgmap v327: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:46 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:46 vm06 ceph-mon[49228]: osdmap e253: 8 total, 8 up, 8 in 2026-03-10T11:36:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:46 vm06 ceph-mon[57405]: pgmap v327: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:46 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:46 vm06 ceph-mon[57405]: osdmap e253: 8 total, 8 up, 8 in 2026-03-10T11:36:46.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:46 vm09 ceph-mon[54793]: pgmap v327: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail 2026-03-10T11:36:46.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:46 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:46.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:46 vm09 ceph-mon[54793]: osdmap e253: 8 total, 8 up, 8 in 2026-03-10T11:36:46.730 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:36:46 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:36:47.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:47 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:47.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:47 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:47.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:47 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:48.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:48 vm09 ceph-mon[54793]: pgmap v330: 164 pgs: 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:48.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:48 vm09 ceph-mon[54793]: osdmap e254: 8 total, 8 up, 8 in 2026-03-10T11:36:48.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:48 vm09 ceph-mon[54793]: osdmap e255: 8 total, 8 up, 8 in 2026-03-10T11:36:48.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:48 vm06 ceph-mon[49228]: pgmap v330: 164 pgs: 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:48.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:48 vm06 ceph-mon[49228]: osdmap e254: 8 total, 8 up, 8 in 2026-03-10T11:36:48.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:48 vm06 ceph-mon[49228]: osdmap e255: 8 total, 8 up, 8 in 2026-03-10T11:36:48.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:48 vm06 ceph-mon[57405]: pgmap v330: 164 pgs: 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:48.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:48 vm06 ceph-mon[57405]: osdmap e254: 8 total, 8 up, 8 in 2026-03-10T11:36:48.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:48 vm06 ceph-mon[57405]: osdmap e255: 8 total, 8 up, 8 in 2026-03-10T11:36:49.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:49 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1886381304' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:49.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:49 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1886381304' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:49.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:49 vm09 ceph-mon[54793]: osdmap e256: 8 total, 8 up, 8 in 2026-03-10T11:36:49.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:49 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1886381304' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:49.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:49 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1886381304' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:49.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:49 vm06 ceph-mon[49228]: osdmap e256: 8 total, 8 up, 8 in 2026-03-10T11:36:49.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:49 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1886381304' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:49.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:49 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1886381304' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:49.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:49 vm06 ceph-mon[57405]: osdmap e256: 8 total, 8 up, 8 in 2026-03-10T11:36:50.286 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_cmpext PASSED [ 68%] 2026-03-10T11:36:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:50 vm09 ceph-mon[54793]: pgmap v333: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:50.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:50 vm09 ceph-mon[54793]: osdmap e257: 8 total, 8 up, 8 in 2026-03-10T11:36:50.743 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:50 vm06 ceph-mon[49228]: pgmap v333: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:50.743 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:50 vm06 ceph-mon[49228]: osdmap e257: 8 total, 8 up, 8 in 2026-03-10T11:36:50.743 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:50 vm06 ceph-mon[57405]: pgmap v333: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 479 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:50.743 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:50 vm06 ceph-mon[57405]: osdmap e257: 8 total, 8 up, 8 in 2026-03-10T11:36:51.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:36:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:36:50] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:36:52.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:52 vm09 ceph-mon[54793]: pgmap v336: 164 pgs: 164 active+clean; 455 KiB data, 480 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:52.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:52 vm09 ceph-mon[54793]: osdmap e258: 8 total, 8 up, 8 in 2026-03-10T11:36:52.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:52 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:52.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:52 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:36:52.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:52 vm06 ceph-mon[49228]: pgmap v336: 164 pgs: 164 active+clean; 455 KiB data, 480 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:52 vm06 ceph-mon[49228]: osdmap e258: 8 total, 8 up, 8 in 2026-03-10T11:36:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:52 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:52.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:52 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:36:52.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:52 vm06 ceph-mon[57405]: pgmap v336: 164 pgs: 164 active+clean; 455 KiB data, 480 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:52.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:52 vm06 ceph-mon[57405]: osdmap e258: 8 total, 8 up, 8 in 2026-03-10T11:36:52.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:52 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:52.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:52 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:36:53.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:53 vm09 ceph-mon[54793]: osdmap e259: 8 total, 8 up, 8 in 2026-03-10T11:36:53.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:53 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2597172864' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:53.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:53 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:53.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:53 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:36:53.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:53 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:36:53.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:53 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[49228]: osdmap e259: 8 total, 8 up, 8 in 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2597172864' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[57405]: osdmap e259: 8 total, 8 up, 8 in 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2597172864' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:36:53.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:53 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:36:54.341 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_rmxattr PASSED [ 69%] 2026-03-10T11:36:54.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:54 vm09 ceph-mon[54793]: pgmap v339: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 480 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:54.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:54 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:54.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:54 vm09 ceph-mon[54793]: osdmap e260: 8 total, 8 up, 8 in 2026-03-10T11:36:54.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:54 vm06 ceph-mon[49228]: pgmap v339: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 480 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:54.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:54 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:54.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:54 vm06 ceph-mon[49228]: osdmap e260: 8 total, 8 up, 8 in 2026-03-10T11:36:54.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:54 vm06 ceph-mon[57405]: pgmap v339: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 480 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:54.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:54 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:54.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:54 vm06 ceph-mon[57405]: osdmap e260: 8 total, 8 up, 8 in 2026-03-10T11:36:55.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:55 vm09 ceph-mon[54793]: osdmap e261: 8 total, 8 up, 8 in 2026-03-10T11:36:55.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:55 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:55.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:55 vm06 ceph-mon[49228]: osdmap e261: 8 total, 8 up, 8 in 2026-03-10T11:36:55.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:55 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:55.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:55 vm06 ceph-mon[57405]: osdmap e261: 8 total, 8 up, 8 in 2026-03-10T11:36:55.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:55 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:36:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:56 vm09 ceph-mon[54793]: pgmap v342: 164 pgs: 164 active+clean; 455 KiB data, 480 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:36:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:56 vm09 ceph-mon[54793]: osdmap e262: 8 total, 8 up, 8 in 2026-03-10T11:36:56.730 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:36:56 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:36:56.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:56 vm06 ceph-mon[49228]: pgmap v342: 164 pgs: 164 active+clean; 455 KiB data, 480 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:36:56.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:56 vm06 ceph-mon[49228]: osdmap e262: 8 total, 8 up, 8 in 2026-03-10T11:36:56.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:56 vm06 ceph-mon[57405]: pgmap v342: 164 pgs: 164 active+clean; 455 KiB data, 480 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:36:56.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:56 vm06 ceph-mon[57405]: osdmap e262: 8 total, 8 up, 8 in 2026-03-10T11:36:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:57 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:57 vm09 ceph-mon[54793]: osdmap e263: 8 total, 8 up, 8 in 2026-03-10T11:36:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:57 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/4089351623' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:57 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:57 vm09 ceph-mon[54793]: pgmap v345: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:57 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:57 vm09 ceph-mon[54793]: osdmap e264: 8 total, 8 up, 8 in 2026-03-10T11:36:57.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[49228]: osdmap e263: 8 total, 8 up, 8 in 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/4089351623' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[49228]: pgmap v345: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[49228]: osdmap e264: 8 total, 8 up, 8 in 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[57405]: osdmap e263: 8 total, 8 up, 8 in 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/4089351623' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[57405]: pgmap v345: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:36:57.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:57 vm06 ceph-mon[57405]: osdmap e264: 8 total, 8 up, 8 in 2026-03-10T11:36:58.434 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_write_no_comp_ref PASSED [ 70%] 2026-03-10T11:36:58.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:58 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:58.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:58 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:58.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:58 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:36:59.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:59 vm09 ceph-mon[54793]: osdmap e265: 8 total, 8 up, 8 in 2026-03-10T11:36:59.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:36:59 vm09 ceph-mon[54793]: pgmap v348: 164 pgs: 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:59.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:59 vm06 ceph-mon[49228]: osdmap e265: 8 total, 8 up, 8 in 2026-03-10T11:36:59.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:36:59 vm06 ceph-mon[49228]: pgmap v348: 164 pgs: 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:36:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:59 vm06 ceph-mon[57405]: osdmap e265: 8 total, 8 up, 8 in 2026-03-10T11:36:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:36:59 vm06 ceph-mon[57405]: pgmap v348: 164 pgs: 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:00 vm09 ceph-mon[54793]: osdmap e266: 8 total, 8 up, 8 in 2026-03-10T11:37:00.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:00 vm06 ceph-mon[49228]: osdmap e266: 8 total, 8 up, 8 in 2026-03-10T11:37:00.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:00 vm06 ceph-mon[57405]: osdmap e266: 8 total, 8 up, 8 in 2026-03-10T11:37:01.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:37:00 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:37:00] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:37:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:01 vm09 ceph-mon[54793]: osdmap e267: 8 total, 8 up, 8 in 2026-03-10T11:37:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:01 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2374470384' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:01 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:01 vm09 ceph-mon[54793]: pgmap v351: 196 pgs: 196 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.5 KiB/s rd, 511 B/s wr, 2 op/s 2026-03-10T11:37:01.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:01 vm06 ceph-mon[49228]: osdmap e267: 8 total, 8 up, 8 in 2026-03-10T11:37:01.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:01 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2374470384' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:01.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:01 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:01.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:01 vm06 ceph-mon[49228]: pgmap v351: 196 pgs: 196 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.5 KiB/s rd, 511 B/s wr, 2 op/s 2026-03-10T11:37:01.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:01 vm06 ceph-mon[57405]: osdmap e267: 8 total, 8 up, 8 in 2026-03-10T11:37:01.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:01 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2374470384' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:01.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:01 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:01.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:01 vm06 ceph-mon[57405]: pgmap v351: 196 pgs: 196 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.5 KiB/s rd, 511 B/s wr, 2 op/s 2026-03-10T11:37:02.481 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_append PASSED [ 71%] 2026-03-10T11:37:02.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:02 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:02.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:02 vm06 ceph-mon[49228]: osdmap e268: 8 total, 8 up, 8 in 2026-03-10T11:37:02.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:02 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:02.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:02 vm06 ceph-mon[57405]: osdmap e268: 8 total, 8 up, 8 in 2026-03-10T11:37:02.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:02 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:02.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:02 vm09 ceph-mon[54793]: osdmap e268: 8 total, 8 up, 8 in 2026-03-10T11:37:03.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:03 vm06 ceph-mon[49228]: osdmap e269: 8 total, 8 up, 8 in 2026-03-10T11:37:03.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:03 vm06 ceph-mon[49228]: pgmap v354: 164 pgs: 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:03.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:03 vm06 ceph-mon[57405]: osdmap e269: 8 total, 8 up, 8 in 2026-03-10T11:37:03.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:03 vm06 ceph-mon[57405]: pgmap v354: 164 pgs: 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:03 vm09 ceph-mon[54793]: osdmap e269: 8 total, 8 up, 8 in 2026-03-10T11:37:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:03 vm09 ceph-mon[54793]: pgmap v354: 164 pgs: 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:04.780 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:04 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:04.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:04 vm06 ceph-mon[49228]: osdmap e270: 8 total, 8 up, 8 in 2026-03-10T11:37:04.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:04 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:04.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:04 vm06 ceph-mon[57405]: osdmap e270: 8 total, 8 up, 8 in 2026-03-10T11:37:04.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:04 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:04.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:04 vm09 ceph-mon[54793]: osdmap e270: 8 total, 8 up, 8 in 2026-03-10T11:37:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:05 vm06 ceph-mon[49228]: osdmap e271: 8 total, 8 up, 8 in 2026-03-10T11:37:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:05 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2486187799' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:05 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:05 vm06 ceph-mon[49228]: pgmap v357: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:05 vm06 ceph-mon[57405]: osdmap e271: 8 total, 8 up, 8 in 2026-03-10T11:37:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:05 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2486187799' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:05 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:05 vm06 ceph-mon[57405]: pgmap v357: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:05 vm09 ceph-mon[54793]: osdmap e271: 8 total, 8 up, 8 in 2026-03-10T11:37:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:05 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2486187799' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:05 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:05 vm09 ceph-mon[54793]: pgmap v357: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 481 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:06.534 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_write_full PASSED [ 72%] 2026-03-10T11:37:06.541 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:37:06 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:37:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:06 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:06 vm09 ceph-mon[54793]: osdmap e272: 8 total, 8 up, 8 in 2026-03-10T11:37:06.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:06 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:07.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:06 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:07.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:06 vm06 ceph-mon[49228]: osdmap e272: 8 total, 8 up, 8 in 2026-03-10T11:37:07.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:06 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:07.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:06 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:07.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:06 vm06 ceph-mon[57405]: osdmap e272: 8 total, 8 up, 8 in 2026-03-10T11:37:07.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:06 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:07 vm09 ceph-mon[54793]: osdmap e273: 8 total, 8 up, 8 in 2026-03-10T11:37:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:07 vm09 ceph-mon[54793]: pgmap v360: 164 pgs: 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:08.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:07 vm06 ceph-mon[49228]: osdmap e273: 8 total, 8 up, 8 in 2026-03-10T11:37:08.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:07 vm06 ceph-mon[49228]: pgmap v360: 164 pgs: 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:08.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:07 vm06 ceph-mon[57405]: osdmap e273: 8 total, 8 up, 8 in 2026-03-10T11:37:08.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:07 vm06 ceph-mon[57405]: pgmap v360: 164 pgs: 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:08.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:08 vm09 ceph-mon[54793]: osdmap e274: 8 total, 8 up, 8 in 2026-03-10T11:37:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:08 vm06 ceph-mon[49228]: osdmap e274: 8 total, 8 up, 8 in 2026-03-10T11:37:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:08 vm06 ceph-mon[57405]: osdmap e274: 8 total, 8 up, 8 in 2026-03-10T11:37:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:09 vm09 ceph-mon[54793]: osdmap e275: 8 total, 8 up, 8 in 2026-03-10T11:37:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:09 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2526333631' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:09 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:09 vm09 ceph-mon[54793]: pgmap v363: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:09 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:09 vm09 ceph-mon[54793]: osdmap e276: 8 total, 8 up, 8 in 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[49228]: osdmap e275: 8 total, 8 up, 8 in 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2526333631' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[49228]: pgmap v363: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[49228]: osdmap e276: 8 total, 8 up, 8 in 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[57405]: osdmap e275: 8 total, 8 up, 8 in 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2526333631' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[57405]: pgmap v363: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:09 vm06 ceph-mon[57405]: osdmap e276: 8 total, 8 up, 8 in 2026-03-10T11:37:10.623 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_writesame PASSED [ 73%] 2026-03-10T11:37:10.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:10 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:10.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:10 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:10.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:10 vm09 ceph-mon[54793]: osdmap e277: 8 total, 8 up, 8 in 2026-03-10T11:37:11.030 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:37:10 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:37:10] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:37:11.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:10 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:11.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:10 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:11.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:10 vm06 ceph-mon[49228]: osdmap e277: 8 total, 8 up, 8 in 2026-03-10T11:37:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:10 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:10 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:10 vm06 ceph-mon[57405]: osdmap e277: 8 total, 8 up, 8 in 2026-03-10T11:37:11.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:11 vm09 ceph-mon[54793]: pgmap v366: 164 pgs: 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:11.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:11 vm09 ceph-mon[54793]: osdmap e278: 8 total, 8 up, 8 in 2026-03-10T11:37:12.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:11 vm06 ceph-mon[49228]: pgmap v366: 164 pgs: 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:11 vm06 ceph-mon[49228]: osdmap e278: 8 total, 8 up, 8 in 2026-03-10T11:37:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:11 vm06 ceph-mon[57405]: pgmap v366: 164 pgs: 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:11 vm06 ceph-mon[57405]: osdmap e278: 8 total, 8 up, 8 in 2026-03-10T11:37:13.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:13 vm09 ceph-mon[54793]: osdmap e279: 8 total, 8 up, 8 in 2026-03-10T11:37:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:13 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1259275119' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:13 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:13 vm09 ceph-mon[54793]: pgmap v369: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:14.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:13 vm06 ceph-mon[49228]: osdmap e279: 8 total, 8 up, 8 in 2026-03-10T11:37:14.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:13 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1259275119' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:14.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:13 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:14.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:13 vm06 ceph-mon[49228]: pgmap v369: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:14.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:13 vm06 ceph-mon[57405]: osdmap e279: 8 total, 8 up, 8 in 2026-03-10T11:37:14.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:13 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1259275119' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:14.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:13 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:14.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:13 vm06 ceph-mon[57405]: pgmap v369: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:14.665 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_stat PASSED [ 74%] 2026-03-10T11:37:14.932 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:14 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:14.932 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:14 vm06 ceph-mon[49228]: osdmap e280: 8 total, 8 up, 8 in 2026-03-10T11:37:14.932 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:14 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:14.932 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:14 vm06 ceph-mon[57405]: osdmap e280: 8 total, 8 up, 8 in 2026-03-10T11:37:14.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:14 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:14.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:14 vm09 ceph-mon[54793]: osdmap e280: 8 total, 8 up, 8 in 2026-03-10T11:37:15.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:15 vm09 ceph-mon[54793]: osdmap e281: 8 total, 8 up, 8 in 2026-03-10T11:37:15.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:15 vm09 ceph-mon[54793]: pgmap v372: 164 pgs: 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:16.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:15 vm06 ceph-mon[49228]: osdmap e281: 8 total, 8 up, 8 in 2026-03-10T11:37:16.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:15 vm06 ceph-mon[49228]: pgmap v372: 164 pgs: 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:16.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:15 vm06 ceph-mon[57405]: osdmap e281: 8 total, 8 up, 8 in 2026-03-10T11:37:16.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:15 vm06 ceph-mon[57405]: pgmap v372: 164 pgs: 164 active+clean; 455 KiB data, 482 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:16.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:37:16 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:37:16.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:16 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:16.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:16 vm09 ceph-mon[54793]: osdmap e282: 8 total, 8 up, 8 in 2026-03-10T11:37:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:16 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:17.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:16 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:16 vm06 ceph-mon[49228]: osdmap e282: 8 total, 8 up, 8 in 2026-03-10T11:37:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:16 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:16 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:16 vm06 ceph-mon[57405]: osdmap e282: 8 total, 8 up, 8 in 2026-03-10T11:37:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:16 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:17.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:17 vm09 ceph-mon[54793]: osdmap e283: 8 total, 8 up, 8 in 2026-03-10T11:37:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:17 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2907736977' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:17 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:17 vm09 ceph-mon[54793]: pgmap v375: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 483 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:17 vm06 ceph-mon[49228]: osdmap e283: 8 total, 8 up, 8 in 2026-03-10T11:37:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:17 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2907736977' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:17 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:17 vm06 ceph-mon[49228]: pgmap v375: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 483 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:17 vm06 ceph-mon[57405]: osdmap e283: 8 total, 8 up, 8 in 2026-03-10T11:37:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:17 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2907736977' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:17 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:17 vm06 ceph-mon[57405]: pgmap v375: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 483 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:18.704 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_remove PASSED [ 75%] 2026-03-10T11:37:18.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:18 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:18.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:18 vm09 ceph-mon[54793]: osdmap e284: 8 total, 8 up, 8 in 2026-03-10T11:37:19.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:18 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:19.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:18 vm06 ceph-mon[49228]: osdmap e284: 8 total, 8 up, 8 in 2026-03-10T11:37:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:18 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:19.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:18 vm06 ceph-mon[57405]: osdmap e284: 8 total, 8 up, 8 in 2026-03-10T11:37:20.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:19 vm06 ceph-mon[49228]: osdmap e285: 8 total, 8 up, 8 in 2026-03-10T11:37:20.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:19 vm06 ceph-mon[49228]: pgmap v378: 164 pgs: 164 active+clean; 455 KiB data, 483 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:20.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:19 vm06 ceph-mon[57405]: osdmap e285: 8 total, 8 up, 8 in 2026-03-10T11:37:20.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:19 vm06 ceph-mon[57405]: pgmap v378: 164 pgs: 164 active+clean; 455 KiB data, 483 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:20.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:19 vm09 ceph-mon[54793]: osdmap e285: 8 total, 8 up, 8 in 2026-03-10T11:37:20.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:19 vm09 ceph-mon[54793]: pgmap v378: 164 pgs: 164 active+clean; 455 KiB data, 483 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:21.030 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:37:20 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:37:20] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:37:21.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:20 vm06 ceph-mon[49228]: osdmap e286: 8 total, 8 up, 8 in 2026-03-10T11:37:21.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:20 vm06 ceph-mon[57405]: osdmap e286: 8 total, 8 up, 8 in 2026-03-10T11:37:21.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:20 vm09 ceph-mon[54793]: osdmap e286: 8 total, 8 up, 8 in 2026-03-10T11:37:22.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:21 vm09 ceph-mon[54793]: osdmap e287: 8 total, 8 up, 8 in 2026-03-10T11:37:22.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:21 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd map", "pool": "test_pool", "object": "foo", "format": "json"}]: dispatch 2026-03-10T11:37:22.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:21 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:22.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:21 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:22.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:21 vm09 ceph-mon[54793]: pgmap v381: 196 pgs: 196 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:37:22.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:21 vm06 ceph-mon[49228]: osdmap e287: 8 total, 8 up, 8 in 2026-03-10T11:37:22.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:21 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd map", "pool": "test_pool", "object": "foo", "format": "json"}]: dispatch 2026-03-10T11:37:22.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:21 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:22.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:21 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:22.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:21 vm06 ceph-mon[49228]: pgmap v381: 196 pgs: 196 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:37:22.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:21 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a[49224]: 2026-03-10T11:37:21.803+0000 7fbaaf32d640 -1 mon.a@0(leader).osd e288 definitely_dead 0 2026-03-10T11:37:22.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:21 vm06 ceph-mon[57405]: osdmap e287: 8 total, 8 up, 8 in 2026-03-10T11:37:22.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:21 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd map", "pool": "test_pool", "object": "foo", "format": "json"}]: dispatch 2026-03-10T11:37:22.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:21 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:22.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:21 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:22.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:21 vm06 ceph-mon[57405]: pgmap v381: 196 pgs: 196 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:37:23.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:22 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:23.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:22 vm09 ceph-mon[54793]: Health check failed: noup flag(s) set (OSDMAP_FLAGS) 2026-03-10T11:37:23.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:22 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd set", "key": "noup"}]': finished 2026-03-10T11:37:23.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:22 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["4", "0", "7"]}]: dispatch 2026-03-10T11:37:23.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:22 vm09 ceph-mon[54793]: osdmap e288: 8 total, 8 up, 8 in 2026-03-10T11:37:23.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:22 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["4", "0", "7"]}]: dispatch 2026-03-10T11:37:23.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:23.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[49228]: Health check failed: noup flag(s) set (OSDMAP_FLAGS) 2026-03-10T11:37:23.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd set", "key": "noup"}]': finished 2026-03-10T11:37:23.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["4", "0", "7"]}]: dispatch 2026-03-10T11:37:23.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[49228]: osdmap e288: 8 total, 8 up, 8 in 2026-03-10T11:37:23.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["4", "0", "7"]}]: dispatch 2026-03-10T11:37:23.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:23.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[57405]: Health check failed: noup flag(s) set (OSDMAP_FLAGS) 2026-03-10T11:37:23.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd set", "key": "noup"}]': finished 2026-03-10T11:37:23.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["4", "0", "7"]}]: dispatch 2026-03-10T11:37:23.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[57405]: osdmap e288: 8 total, 8 up, 8 in 2026-03-10T11:37:23.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:22 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["4", "0", "7"]}]: dispatch 2026-03-10T11:37:23.729 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:37:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4[59293]: 2026-03-10T11:37:23.562+0000 7f530ece6640 -1 osd.4 289 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:23.730 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:37:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:37:23.434+0000 7f5097e22640 -1 osd.7 289 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:23.781 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:37:23 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0[61331]: 2026-03-10T11:37:23.419+0000 7f8b99299640 -1 osd.0 289 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:24.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: Health check failed: 3 osds down (OSD_DOWN) 2026-03-10T11:37:24.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd down", "ids": ["4", "0", "7"]}]': finished 2026-03-10T11:37:24.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: osdmap e289: 8 total, 5 up, 8 in 2026-03-10T11:37:24.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: Monitor daemon marked osd.0 down, but it is still running 2026-03-10T11:37:24.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: map e289 wrongly marked me down at e289 2026-03-10T11:37:24.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: osd.0 marked itself dead as of e289 2026-03-10T11:37:24.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: pgmap v384: 196 pgs: 84 stale+active+clean, 112 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:37:24.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: Monitor daemon marked osd.4 down, but it is still running 2026-03-10T11:37:24.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: map e289 wrongly marked me down at e289 2026-03-10T11:37:24.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: osd.4 marked itself dead as of e289 2026-03-10T11:37:24.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: osd.7 marked itself dead as of e289 2026-03-10T11:37:24.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:23 vm09 ceph-mon[54793]: osdmap e290: 8 total, 5 up, 8 in 2026-03-10T11:37:24.230 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:37:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4[59293]: 2026-03-10T11:37:23.817+0000 7f53020d5640 -1 osd.4 290 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:24.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:37:23 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:37:23.813+0000 7f508a9fe640 -1 osd.7 290 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: Health check failed: 3 osds down (OSD_DOWN) 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd down", "ids": ["4", "0", "7"]}]': finished 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: osdmap e289: 8 total, 5 up, 8 in 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: Monitor daemon marked osd.0 down, but it is still running 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: map e289 wrongly marked me down at e289 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: osd.0 marked itself dead as of e289 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: pgmap v384: 196 pgs: 84 stale+active+clean, 112 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: Monitor daemon marked osd.4 down, but it is still running 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: map e289 wrongly marked me down at e289 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: osd.4 marked itself dead as of e289 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: osd.7 marked itself dead as of e289 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[49228]: osdmap e290: 8 total, 5 up, 8 in 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: Health check failed: 3 osds down (OSD_DOWN) 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd down", "ids": ["4", "0", "7"]}]': finished 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: osdmap e289: 8 total, 5 up, 8 in 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: Monitor daemon marked osd.0 down, but it is still running 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: map e289 wrongly marked me down at e289 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: osd.0 marked itself dead as of e289 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: pgmap v384: 196 pgs: 84 stale+active+clean, 112 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: Monitor daemon marked osd.4 down, but it is still running 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: map e289 wrongly marked me down at e289 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: osd.4 marked itself dead as of e289 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: osd.7 marked itself dead as of e289 2026-03-10T11:37:24.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:23 vm06 ceph-mon[57405]: osdmap e290: 8 total, 5 up, 8 in 2026-03-10T11:37:24.281 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:37:23 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0[61331]: 2026-03-10T11:37:23.824+0000 7f8b8c676640 -1 osd.0 290 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:25.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:24 vm09 ceph-mon[54793]: Monitor daemon marked osd.7 down, but it is still running 2026-03-10T11:37:25.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:24 vm09 ceph-mon[54793]: map e289 wrongly marked me down at e289 2026-03-10T11:37:25.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:24 vm06 ceph-mon[49228]: Monitor daemon marked osd.7 down, but it is still running 2026-03-10T11:37:25.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:24 vm06 ceph-mon[49228]: map e289 wrongly marked me down at e289 2026-03-10T11:37:25.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:24 vm06 ceph-mon[57405]: Monitor daemon marked osd.7 down, but it is still running 2026-03-10T11:37:25.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:24 vm06 ceph-mon[57405]: map e289 wrongly marked me down at e289 2026-03-10T11:37:25.842 INFO:journalctl@ceph.rgw.foo.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-rgw-foo-a[81067]: 2026-03-10T11:37:25.573+0000 7f39f4114640 -1 rgw watcher librados: RGWWatcher::handle_error cookie 94091011383168 err (110) Connection timed out 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: pgmap v386: 196 pgs: 84 stale+active+clean, 112 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 239 B/s wr, 1 op/s 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.a", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.d", "id": [1, 7]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [3, 0, 1, 4]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.14", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.17", "id": [3, 0]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.19", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1b", "id": [3, 7, 1, 0]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.a", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.d", "id": [1, 7]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [3, 0, 1, 4]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.14", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.17", "id": [3, 0]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.19", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1b", "id": [3, 7, 1, 0]}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:25 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:26.230 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:37:25 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4[59293]: 2026-03-10T11:37:25.838+0000 7f530a30f640 -1 osd.4 291 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:37:26.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:37:25 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:37:25.838+0000 7f5092c38640 -1 osd.7 291 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: pgmap v386: 196 pgs: 84 stale+active+clean, 112 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 239 B/s wr, 1 op/s 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.a", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.d", "id": [1, 7]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [3, 0, 1, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.14", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.17", "id": [3, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.19", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1b", "id": [3, 7, 1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.a", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.d", "id": [1, 7]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [3, 0, 1, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.14", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.17", "id": [3, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.19", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1b", "id": [3, 7, 1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: pgmap v386: 196 pgs: 84 stale+active+clean, 112 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 239 B/s wr, 1 op/s 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.a", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.d", "id": [1, 7]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [3, 0, 1, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.14", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.17", "id": [3, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.19", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1b", "id": [3, 7, 1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.a", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.d", "id": [1, 7]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [3, 0, 1, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.14", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.17", "id": [3, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.19", "id": [3, 4]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1b", "id": [3, 7, 1, 0]}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:25 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:26.281 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:37:25 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0[61331]: 2026-03-10T11:37:25.836+0000 7f8b948b0640 -1 osd.0 291 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:37:26.730 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:37:26 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: Health check cleared: OSDMAP_FLAGS (was: noup flag(s) set) 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.a", "id": [1, 0]}]': finished 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.d", "id": [1, 7]}]': finished 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [3, 0, 1, 4]}]': finished 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.14", "id": [3, 4]}]': finished 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.17", "id": [3, 0]}]': finished 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 0]}]': finished 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.19", "id": [3, 4]}]': finished 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1b", "id": [3, 7, 1, 0]}]': finished 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: osdmap e291: 8 total, 5 up, 8 in 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:27.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:26 vm09 ceph-mon[54793]: Health check cleared: OSD_DOWN (was: 3 osds down) 2026-03-10T11:37:27.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: Health check cleared: OSDMAP_FLAGS (was: noup flag(s) set) 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.a", "id": [1, 0]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.d", "id": [1, 7]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [3, 0, 1, 4]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.14", "id": [3, 4]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.17", "id": [3, 0]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 0]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.19", "id": [3, 4]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1b", "id": [3, 7, 1, 0]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: osdmap e291: 8 total, 5 up, 8 in 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[49228]: Health check cleared: OSD_DOWN (was: 3 osds down) 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: Health check cleared: OSDMAP_FLAGS (was: noup flag(s) set) 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.a", "id": [1, 0]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.d", "id": [1, 7]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [3, 0, 1, 4]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.14", "id": [3, 4]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.17", "id": [3, 0]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 0]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.19", "id": [3, 4]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1b", "id": [3, 7, 1, 0]}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: osdmap e291: 8 total, 5 up, 8 in 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:27.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:26 vm06 ceph-mon[57405]: Health check cleared: OSD_DOWN (was: 3 osds down) 2026-03-10T11:37:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:27 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:37:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:27 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:37:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:27 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:37:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:27 vm09 ceph-mon[54793]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:37:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:27 vm09 ceph-mon[54793]: osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642] boot 2026-03-10T11:37:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:27 vm09 ceph-mon[54793]: osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574] boot 2026-03-10T11:37:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:27 vm09 ceph-mon[54793]: osdmap e292: 8 total, 8 up, 8 in 2026-03-10T11:37:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:27 vm09 ceph-mon[54793]: pgmap v389: 196 pgs: 6 unknown, 1 peering, 73 active+undersized, 43 undersized+peered, 5 stale+active+clean, 22 active+undersized+degraded, 14 undersized+degraded+peered, 32 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 178/600 objects degraded (29.667%) 2026-03-10T11:37:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:27 vm09 ceph-mon[54793]: osdmap e293: 8 total, 8 up, 8 in 2026-03-10T11:37:28.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[49228]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[49228]: osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642] boot 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[49228]: osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574] boot 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[49228]: osdmap e292: 8 total, 8 up, 8 in 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[49228]: pgmap v389: 196 pgs: 6 unknown, 1 peering, 73 active+undersized, 43 undersized+peered, 5 stale+active+clean, 22 active+undersized+degraded, 14 undersized+degraded+peered, 32 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 178/600 objects degraded (29.667%) 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[49228]: osdmap e293: 8 total, 8 up, 8 in 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[57405]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[57405]: osd.4 [v2:192.168.123.109:6800/3122480642,v1:192.168.123.109:6801/3122480642] boot 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[57405]: osd.0 [v2:192.168.123.106:6802/3972625574,v1:192.168.123.106:6803/3972625574] boot 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[57405]: osdmap e292: 8 total, 8 up, 8 in 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[57405]: pgmap v389: 196 pgs: 6 unknown, 1 peering, 73 active+undersized, 43 undersized+peered, 5 stale+active+clean, 22 active+undersized+degraded, 14 undersized+degraded+peered, 32 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 178/600 objects degraded (29.667%) 2026-03-10T11:37:28.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:27 vm06 ceph-mon[57405]: osdmap e293: 8 total, 8 up, 8 in 2026-03-10T11:37:29.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:28 vm09 ceph-mon[54793]: Health check failed: Reduced data availability: 25 pgs inactive (PG_AVAILABILITY) 2026-03-10T11:37:29.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:28 vm09 ceph-mon[54793]: Health check failed: Degraded data redundancy: 178/600 objects degraded (29.667%), 36 pgs degraded (PG_DEGRADED) 2026-03-10T11:37:29.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:28 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:29.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:28 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:29.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:28 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:29.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:28 vm09 ceph-mon[54793]: osdmap e294: 8 total, 8 up, 8 in 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[49228]: Health check failed: Reduced data availability: 25 pgs inactive (PG_AVAILABILITY) 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[49228]: Health check failed: Degraded data redundancy: 178/600 objects degraded (29.667%), 36 pgs degraded (PG_DEGRADED) 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[49228]: osdmap e294: 8 total, 8 up, 8 in 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[57405]: Health check failed: Reduced data availability: 25 pgs inactive (PG_AVAILABILITY) 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[57405]: Health check failed: Degraded data redundancy: 178/600 objects degraded (29.667%), 36 pgs degraded (PG_DEGRADED) 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3166663404' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:29.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:28 vm06 ceph-mon[57405]: osdmap e294: 8 total, 8 up, 8 in 2026-03-10T11:37:29.852 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_read_wait_for_complete PASSED [ 76%] 2026-03-10T11:37:30.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:29 vm09 ceph-mon[54793]: pgmap v392: 196 pgs: 6 unknown, 1 peering, 73 active+undersized, 43 undersized+peered, 5 stale+active+clean, 22 active+undersized+degraded, 14 undersized+degraded+peered, 32 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 178/600 objects degraded (29.667%) 2026-03-10T11:37:30.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:29 vm09 ceph-mon[54793]: osdmap e295: 8 total, 8 up, 8 in 2026-03-10T11:37:30.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:29 vm06 ceph-mon[49228]: pgmap v392: 196 pgs: 6 unknown, 1 peering, 73 active+undersized, 43 undersized+peered, 5 stale+active+clean, 22 active+undersized+degraded, 14 undersized+degraded+peered, 32 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 178/600 objects degraded (29.667%) 2026-03-10T11:37:30.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:29 vm06 ceph-mon[49228]: osdmap e295: 8 total, 8 up, 8 in 2026-03-10T11:37:30.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:29 vm06 ceph-mon[57405]: pgmap v392: 196 pgs: 6 unknown, 1 peering, 73 active+undersized, 43 undersized+peered, 5 stale+active+clean, 22 active+undersized+degraded, 14 undersized+degraded+peered, 32 active+clean; 455 KiB data, 484 MiB used, 159 GiB / 160 GiB avail; 178/600 objects degraded (29.667%) 2026-03-10T11:37:30.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:29 vm06 ceph-mon[57405]: osdmap e295: 8 total, 8 up, 8 in 2026-03-10T11:37:31.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:37:30 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:37:30] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:37:32.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:31 vm09 ceph-mon[54793]: osdmap e296: 8 total, 8 up, 8 in 2026-03-10T11:37:32.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:31 vm09 ceph-mon[54793]: pgmap v395: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 485 MiB used, 159 GiB / 160 GiB avail; 84 KiB/s rd, 84 op/s; 137 B/s, 10 objects/s recovering 2026-03-10T11:37:32.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:31 vm06 ceph-mon[49228]: osdmap e296: 8 total, 8 up, 8 in 2026-03-10T11:37:32.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:31 vm06 ceph-mon[49228]: pgmap v395: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 485 MiB used, 159 GiB / 160 GiB avail; 84 KiB/s rd, 84 op/s; 137 B/s, 10 objects/s recovering 2026-03-10T11:37:32.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:31 vm06 ceph-mon[57405]: osdmap e296: 8 total, 8 up, 8 in 2026-03-10T11:37:32.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:31 vm06 ceph-mon[57405]: pgmap v395: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 485 MiB used, 159 GiB / 160 GiB avail; 84 KiB/s rd, 84 op/s; 137 B/s, 10 objects/s recovering 2026-03-10T11:37:33.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:32 vm09 ceph-mon[54793]: Health check cleared: PG_AVAILABILITY (was: Reduced data availability: 25 pgs inactive) 2026-03-10T11:37:33.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:32 vm09 ceph-mon[54793]: Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 178/600 objects degraded (29.667%), 36 pgs degraded) 2026-03-10T11:37:33.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:32 vm09 ceph-mon[54793]: osdmap e297: 8 total, 8 up, 8 in 2026-03-10T11:37:33.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:32 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd map", "pool": "test_pool", "object": "foo", "format": "json"}]: dispatch 2026-03-10T11:37:33.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:32 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:33.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:32 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a[49224]: 2026-03-10T11:37:32.862+0000 7fbaaf32d640 -1 mon.a@0(leader).osd e298 definitely_dead 0 2026-03-10T11:37:33.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:32 vm06 ceph-mon[49228]: Health check cleared: PG_AVAILABILITY (was: Reduced data availability: 25 pgs inactive) 2026-03-10T11:37:33.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:32 vm06 ceph-mon[49228]: Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 178/600 objects degraded (29.667%), 36 pgs degraded) 2026-03-10T11:37:33.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:32 vm06 ceph-mon[49228]: osdmap e297: 8 total, 8 up, 8 in 2026-03-10T11:37:33.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:32 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd map", "pool": "test_pool", "object": "foo", "format": "json"}]: dispatch 2026-03-10T11:37:33.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:32 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:33.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:32 vm06 ceph-mon[57405]: Health check cleared: PG_AVAILABILITY (was: Reduced data availability: 25 pgs inactive) 2026-03-10T11:37:33.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:32 vm06 ceph-mon[57405]: Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 178/600 objects degraded (29.667%), 36 pgs degraded) 2026-03-10T11:37:33.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:32 vm06 ceph-mon[57405]: osdmap e297: 8 total, 8 up, 8 in 2026-03-10T11:37:33.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:32 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd map", "pool": "test_pool", "object": "foo", "format": "json"}]: dispatch 2026-03-10T11:37:33.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:32 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:34.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:33 vm09 ceph-mon[54793]: Health check failed: noup flag(s) set (OSDMAP_FLAGS) 2026-03-10T11:37:34.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:33 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd set", "key": "noup"}]': finished 2026-03-10T11:37:34.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:33 vm09 ceph-mon[54793]: osdmap e298: 8 total, 8 up, 8 in 2026-03-10T11:37:34.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:33 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["2", "5", "7"]}]: dispatch 2026-03-10T11:37:34.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:33 vm09 ceph-mon[54793]: pgmap v398: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 485 MiB used, 159 GiB / 160 GiB avail; 84 KiB/s rd, 84 op/s; 137 B/s, 10 objects/s recovering 2026-03-10T11:37:34.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:33 vm09 ceph-mon[54793]: Health check failed: 3 osds down (OSD_DOWN) 2026-03-10T11:37:34.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[49228]: Health check failed: noup flag(s) set (OSDMAP_FLAGS) 2026-03-10T11:37:34.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd set", "key": "noup"}]': finished 2026-03-10T11:37:34.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[49228]: osdmap e298: 8 total, 8 up, 8 in 2026-03-10T11:37:34.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["2", "5", "7"]}]: dispatch 2026-03-10T11:37:34.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[49228]: pgmap v398: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 485 MiB used, 159 GiB / 160 GiB avail; 84 KiB/s rd, 84 op/s; 137 B/s, 10 objects/s recovering 2026-03-10T11:37:34.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[49228]: Health check failed: 3 osds down (OSD_DOWN) 2026-03-10T11:37:34.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[57405]: Health check failed: noup flag(s) set (OSDMAP_FLAGS) 2026-03-10T11:37:34.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd set", "key": "noup"}]': finished 2026-03-10T11:37:34.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[57405]: osdmap e298: 8 total, 8 up, 8 in 2026-03-10T11:37:34.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["2", "5", "7"]}]: dispatch 2026-03-10T11:37:34.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[57405]: pgmap v398: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 485 MiB used, 159 GiB / 160 GiB avail; 84 KiB/s rd, 84 op/s; 137 B/s, 10 objects/s recovering 2026-03-10T11:37:34.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:33 vm06 ceph-mon[57405]: Health check failed: 3 osds down (OSD_DOWN) 2026-03-10T11:37:34.931 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:37:34 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:37:34.585+0000 7f509760f640 -1 osd.7 299 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:35.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:34 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd down", "ids": ["2", "5", "7"]}]': finished 2026-03-10T11:37:35.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:34 vm09 ceph-mon[54793]: osdmap e299: 8 total, 5 up, 8 in 2026-03-10T11:37:35.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:34 vm09 ceph-mon[54793]: osd.2 marked itself dead as of e299 2026-03-10T11:37:35.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:34 vm09 ceph-mon[54793]: osd.7 marked itself dead as of e299 2026-03-10T11:37:35.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:37:34 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:37:34.928+0000 7f508a9fe640 -1 osd.7 300 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:35.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:34 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd down", "ids": ["2", "5", "7"]}]': finished 2026-03-10T11:37:35.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:34 vm06 ceph-mon[49228]: osdmap e299: 8 total, 5 up, 8 in 2026-03-10T11:37:35.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:34 vm06 ceph-mon[49228]: osd.2 marked itself dead as of e299 2026-03-10T11:37:35.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:34 vm06 ceph-mon[49228]: osd.7 marked itself dead as of e299 2026-03-10T11:37:35.281 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:37:34 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2[71586]: 2026-03-10T11:37:34.911+0000 7f4b56f2d640 -1 osd.2 299 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:35.281 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:37:34 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2[71586]: 2026-03-10T11:37:34.938+0000 7f4b4a31c640 -1 osd.2 300 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:35.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:34 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd down", "ids": ["2", "5", "7"]}]': finished 2026-03-10T11:37:35.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:34 vm06 ceph-mon[57405]: osdmap e299: 8 total, 5 up, 8 in 2026-03-10T11:37:35.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:34 vm06 ceph-mon[57405]: osd.2 marked itself dead as of e299 2026-03-10T11:37:35.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:34 vm06 ceph-mon[57405]: osd.7 marked itself dead as of e299 2026-03-10T11:37:36.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:35 vm09 ceph-mon[54793]: Monitor daemon marked osd.2 down, but it is still running 2026-03-10T11:37:36.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:35 vm09 ceph-mon[54793]: map e299 wrongly marked me down at e299 2026-03-10T11:37:36.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:35 vm09 ceph-mon[54793]: Monitor daemon marked osd.7 down, but it is still running 2026-03-10T11:37:36.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:35 vm09 ceph-mon[54793]: map e299 wrongly marked me down at e299 2026-03-10T11:37:36.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:35 vm09 ceph-mon[54793]: osdmap e300: 8 total, 5 up, 8 in 2026-03-10T11:37:36.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:35 vm09 ceph-mon[54793]: pgmap v401: 196 pgs: 16 stale+creating+peering, 54 stale+active+clean, 16 creating+peering, 110 active+clean; 455 KiB data, 485 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:36.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:35 vm09 ceph-mon[54793]: osd.5 marked itself dead as of e300 2026-03-10T11:37:36.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:37:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:37:35.951+0000 7f508a9fe640 -1 osd.7 301 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:36.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[49228]: Monitor daemon marked osd.2 down, but it is still running 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[49228]: map e299 wrongly marked me down at e299 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[49228]: Monitor daemon marked osd.7 down, but it is still running 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[49228]: map e299 wrongly marked me down at e299 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[49228]: osdmap e300: 8 total, 5 up, 8 in 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[49228]: pgmap v401: 196 pgs: 16 stale+creating+peering, 54 stale+active+clean, 16 creating+peering, 110 active+clean; 455 KiB data, 485 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[49228]: osd.5 marked itself dead as of e300 2026-03-10T11:37:36.281 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:37:35 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2[71586]: 2026-03-10T11:37:35.951+0000 7f4b4a31c640 -1 osd.2 301 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[57405]: Monitor daemon marked osd.2 down, but it is still running 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[57405]: map e299 wrongly marked me down at e299 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[57405]: Monitor daemon marked osd.7 down, but it is still running 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[57405]: map e299 wrongly marked me down at e299 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[57405]: osdmap e300: 8 total, 5 up, 8 in 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[57405]: pgmap v401: 196 pgs: 16 stale+creating+peering, 54 stale+active+clean, 16 creating+peering, 110 active+clean; 455 KiB data, 485 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:35 vm06 ceph-mon[57405]: osd.5 marked itself dead as of e300 2026-03-10T11:37:36.657 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:37:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:37:36.951 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:37:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:37:36.657+0000 7f578a08a640 -1 osd.5 301 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:37.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:36 vm09 ceph-mon[54793]: Monitor daemon marked osd.5 down, but it is still running 2026-03-10T11:37:37.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:36 vm09 ceph-mon[54793]: map e300 wrongly marked me down at e299 2026-03-10T11:37:37.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:36 vm09 ceph-mon[54793]: osdmap e301: 8 total, 5 up, 8 in 2026-03-10T11:37:37.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:36 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:37.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:36 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:37.230 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:37:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:37:36.967+0000 7f5785eb4640 -1 osd.5 302 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:37:37.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:37:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:37:36.960+0000 7f5092c38640 -1 osd.7 302 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:37:37.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:36 vm06 ceph-mon[49228]: Monitor daemon marked osd.5 down, but it is still running 2026-03-10T11:37:37.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:36 vm06 ceph-mon[49228]: map e300 wrongly marked me down at e299 2026-03-10T11:37:37.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:36 vm06 ceph-mon[49228]: osdmap e301: 8 total, 5 up, 8 in 2026-03-10T11:37:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:36 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:36 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:37.281 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:37:36 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2[71586]: 2026-03-10T11:37:36.959+0000 7f4b52556640 -1 osd.2 302 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:37:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:36 vm06 ceph-mon[57405]: Monitor daemon marked osd.5 down, but it is still running 2026-03-10T11:37:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:36 vm06 ceph-mon[57405]: map e300 wrongly marked me down at e299 2026-03-10T11:37:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:36 vm06 ceph-mon[57405]: osdmap e301: 8 total, 5 up, 8 in 2026-03-10T11:37:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:36 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:36 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:37 vm06 ceph-mon[49228]: Health check cleared: OSDMAP_FLAGS (was: noup flag(s) set) 2026-03-10T11:37:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:37 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:37 vm06 ceph-mon[49228]: osdmap e302: 8 total, 5 up, 8 in 2026-03-10T11:37:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:37 vm06 ceph-mon[49228]: pgmap v404: 196 pgs: 2 undersized+degraded+peered+wait, 3 active+undersized+degraded+wait, 25 undersized+peered, 70 active+undersized, 2 stale+creating+peering, 5 stale+active+clean, 6 undersized+degraded+peered, 6 undersized+peered+wait, 16 active+undersized+wait, 25 active+undersized+degraded, 36 active+clean; 455 KiB data, 494 MiB used, 159 GiB / 160 GiB avail; 191/597 objects degraded (31.993%) 2026-03-10T11:37:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:37 vm06 ceph-mon[57405]: Health check cleared: OSDMAP_FLAGS (was: noup flag(s) set) 2026-03-10T11:37:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:37 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:37 vm06 ceph-mon[57405]: osdmap e302: 8 total, 5 up, 8 in 2026-03-10T11:37:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:37 vm06 ceph-mon[57405]: pgmap v404: 196 pgs: 2 undersized+degraded+peered+wait, 3 active+undersized+degraded+wait, 25 undersized+peered, 70 active+undersized, 2 stale+creating+peering, 5 stale+active+clean, 6 undersized+degraded+peered, 6 undersized+peered+wait, 16 active+undersized+wait, 25 active+undersized+degraded, 36 active+clean; 455 KiB data, 494 MiB used, 159 GiB / 160 GiB avail; 191/597 objects degraded (31.993%) 2026-03-10T11:37:38.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:37 vm09 ceph-mon[54793]: Health check cleared: OSDMAP_FLAGS (was: noup flag(s) set) 2026-03-10T11:37:38.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:37 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:38.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:37 vm09 ceph-mon[54793]: osdmap e302: 8 total, 5 up, 8 in 2026-03-10T11:37:38.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:37 vm09 ceph-mon[54793]: pgmap v404: 196 pgs: 2 undersized+degraded+peered+wait, 3 active+undersized+degraded+wait, 25 undersized+peered, 70 active+undersized, 2 stale+creating+peering, 5 stale+active+clean, 6 undersized+degraded+peered, 6 undersized+peered+wait, 16 active+undersized+wait, 25 active+undersized+degraded, 36 active+clean; 455 KiB data, 494 MiB used, 159 GiB / 160 GiB avail; 191/597 objects degraded (31.993%) 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[49228]: Health check cleared: OSD_DOWN (was: 3 osds down) 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[49228]: Health check failed: Reduced data availability: 2 pgs inactive (PG_AVAILABILITY) 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[49228]: Health check failed: Degraded data redundancy: 191/597 objects degraded (31.993%), 36 pgs degraded (PG_DEGRADED) 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[49228]: osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931] boot 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[49228]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[49228]: osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179] boot 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[49228]: osdmap e303: 8 total, 8 up, 8 in 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[57405]: Health check cleared: OSD_DOWN (was: 3 osds down) 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[57405]: Health check failed: Reduced data availability: 2 pgs inactive (PG_AVAILABILITY) 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[57405]: Health check failed: Degraded data redundancy: 191/597 objects degraded (31.993%), 36 pgs degraded (PG_DEGRADED) 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[57405]: osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931] boot 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[57405]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[57405]: osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179] boot 2026-03-10T11:37:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:38 vm06 ceph-mon[57405]: osdmap e303: 8 total, 8 up, 8 in 2026-03-10T11:37:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:38 vm09 ceph-mon[54793]: Health check cleared: OSD_DOWN (was: 3 osds down) 2026-03-10T11:37:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:38 vm09 ceph-mon[54793]: Health check failed: Reduced data availability: 2 pgs inactive (PG_AVAILABILITY) 2026-03-10T11:37:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:38 vm09 ceph-mon[54793]: Health check failed: Degraded data redundancy: 191/597 objects degraded (31.993%), 36 pgs degraded (PG_DEGRADED) 2026-03-10T11:37:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:38 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:37:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:38 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T11:37:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:38 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:37:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:38 vm09 ceph-mon[54793]: osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931] boot 2026-03-10T11:37:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:38 vm09 ceph-mon[54793]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:37:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:38 vm09 ceph-mon[54793]: osd.5 [v2:192.168.123.109:6808/741270179,v1:192.168.123.109:6809/741270179] boot 2026-03-10T11:37:39.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:38 vm09 ceph-mon[54793]: osdmap e303: 8 total, 8 up, 8 in 2026-03-10T11:37:40.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:40 vm06 ceph-mon[49228]: osdmap e304: 8 total, 8 up, 8 in 2026-03-10T11:37:40.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:40 vm06 ceph-mon[49228]: pgmap v407: 196 pgs: 2 undersized+degraded+peered+wait, 3 active+undersized+degraded+wait, 25 undersized+peered, 70 active+undersized, 2 stale+creating+peering, 5 stale+active+clean, 6 undersized+degraded+peered, 6 undersized+peered+wait, 16 active+undersized+wait, 25 active+undersized+degraded, 36 active+clean; 455 KiB data, 494 MiB used, 159 GiB / 160 GiB avail; 191/597 objects degraded (31.993%) 2026-03-10T11:37:40.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:40 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:40.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:40 vm06 ceph-mon[57405]: osdmap e304: 8 total, 8 up, 8 in 2026-03-10T11:37:40.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:40 vm06 ceph-mon[57405]: pgmap v407: 196 pgs: 2 undersized+degraded+peered+wait, 3 active+undersized+degraded+wait, 25 undersized+peered, 70 active+undersized, 2 stale+creating+peering, 5 stale+active+clean, 6 undersized+degraded+peered, 6 undersized+peered+wait, 16 active+undersized+wait, 25 active+undersized+degraded, 36 active+clean; 455 KiB data, 494 MiB used, 159 GiB / 160 GiB avail; 191/597 objects degraded (31.993%) 2026-03-10T11:37:40.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:40 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:40.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:40 vm09 ceph-mon[54793]: osdmap e304: 8 total, 8 up, 8 in 2026-03-10T11:37:40.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:40 vm09 ceph-mon[54793]: pgmap v407: 196 pgs: 2 undersized+degraded+peered+wait, 3 active+undersized+degraded+wait, 25 undersized+peered, 70 active+undersized, 2 stale+creating+peering, 5 stale+active+clean, 6 undersized+degraded+peered, 6 undersized+peered+wait, 16 active+undersized+wait, 25 active+undersized+degraded, 36 active+clean; 455 KiB data, 494 MiB used, 159 GiB / 160 GiB avail; 191/597 objects degraded (31.993%) 2026-03-10T11:37:40.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:40 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:41.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:37:40 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:37:40] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:37:41.043 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_read_wait_for_complete_and_cb PASSED [ 78%] 2026-03-10T11:37:41.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:41 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:41.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:41 vm09 ceph-mon[54793]: osdmap e305: 8 total, 8 up, 8 in 2026-03-10T11:37:41.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:41 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:41.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:41 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:41.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:41 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:41.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:41 vm06 ceph-mon[49228]: osdmap e305: 8 total, 8 up, 8 in 2026-03-10T11:37:41.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:41 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:41.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:41 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:41.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:41 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3585483967' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:41.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:41 vm06 ceph-mon[57405]: osdmap e305: 8 total, 8 up, 8 in 2026-03-10T11:37:41.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:41 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:41.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:41 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:42.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:42 vm09 ceph-mon[54793]: osdmap e306: 8 total, 8 up, 8 in 2026-03-10T11:37:42.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:42 vm09 ceph-mon[54793]: pgmap v410: 164 pgs: 164 active+clean; 455 KiB data, 498 MiB used, 159 GiB / 160 GiB avail; 767 B/s rd, 0 op/s; 82 B/s, 2 objects/s recovering 2026-03-10T11:37:42.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:42 vm06 ceph-mon[49228]: osdmap e306: 8 total, 8 up, 8 in 2026-03-10T11:37:42.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:42 vm06 ceph-mon[49228]: pgmap v410: 164 pgs: 164 active+clean; 455 KiB data, 498 MiB used, 159 GiB / 160 GiB avail; 767 B/s rd, 0 op/s; 82 B/s, 2 objects/s recovering 2026-03-10T11:37:42.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:42 vm06 ceph-mon[57405]: osdmap e306: 8 total, 8 up, 8 in 2026-03-10T11:37:42.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:42 vm06 ceph-mon[57405]: pgmap v410: 164 pgs: 164 active+clean; 455 KiB data, 498 MiB used, 159 GiB / 160 GiB avail; 767 B/s rd, 0 op/s; 82 B/s, 2 objects/s recovering 2026-03-10T11:37:43.381 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:43 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:43.381 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:43 vm09 ceph-mon[54793]: Health check cleared: PG_AVAILABILITY (was: Reduced data availability: 2 pgs inactive) 2026-03-10T11:37:43.381 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:43 vm09 ceph-mon[54793]: Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 191/597 objects degraded (31.993%), 36 pgs degraded) 2026-03-10T11:37:43.381 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:43 vm09 ceph-mon[54793]: osdmap e307: 8 total, 8 up, 8 in 2026-03-10T11:37:43.381 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:43 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd map", "pool": "test_pool", "object": "bar", "format": "json"}]: dispatch 2026-03-10T11:37:43.381 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:43 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:43.381 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:43 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[49228]: Health check cleared: PG_AVAILABILITY (was: Reduced data availability: 2 pgs inactive) 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[49228]: Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 191/597 objects degraded (31.993%), 36 pgs degraded) 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[49228]: osdmap e307: 8 total, 8 up, 8 in 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd map", "pool": "test_pool", "object": "bar", "format": "json"}]: dispatch 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:43 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a[49224]: 2026-03-10T11:37:43.120+0000 7fbaaf32d640 -1 mon.a@0(leader).osd e308 definitely_dead 0 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[57405]: Health check cleared: PG_AVAILABILITY (was: Reduced data availability: 2 pgs inactive) 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[57405]: Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 191/597 objects degraded (31.993%), 36 pgs degraded) 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[57405]: osdmap e307: 8 total, 8 up, 8 in 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd map", "pool": "test_pool", "object": "bar", "format": "json"}]: dispatch 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:43.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:43 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd set", "key": "noup"}]: dispatch 2026-03-10T11:37:44.432 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[49228]: pgmap v412: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 498 MiB used, 159 GiB / 160 GiB avail; 754 B/s rd, 0 op/s; 81 B/s, 2 objects/s recovering 2026-03-10T11:37:44.433 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[49228]: Health check failed: noup flag(s) set (OSDMAP_FLAGS) 2026-03-10T11:37:44.433 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd set", "key": "noup"}]': finished 2026-03-10T11:37:44.433 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["1", "7", "2"]}]: dispatch 2026-03-10T11:37:44.433 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[49228]: osdmap e308: 8 total, 8 up, 8 in 2026-03-10T11:37:44.433 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["1", "7", "2"]}]: dispatch 2026-03-10T11:37:44.433 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[57405]: pgmap v412: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 498 MiB used, 159 GiB / 160 GiB avail; 754 B/s rd, 0 op/s; 81 B/s, 2 objects/s recovering 2026-03-10T11:37:44.433 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[57405]: Health check failed: noup flag(s) set (OSDMAP_FLAGS) 2026-03-10T11:37:44.433 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd set", "key": "noup"}]': finished 2026-03-10T11:37:44.433 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["1", "7", "2"]}]: dispatch 2026-03-10T11:37:44.433 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[57405]: osdmap e308: 8 total, 8 up, 8 in 2026-03-10T11:37:44.433 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:44 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["1", "7", "2"]}]: dispatch 2026-03-10T11:37:44.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:44 vm09 ceph-mon[54793]: pgmap v412: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 498 MiB used, 159 GiB / 160 GiB avail; 754 B/s rd, 0 op/s; 81 B/s, 2 objects/s recovering 2026-03-10T11:37:44.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:44 vm09 ceph-mon[54793]: Health check failed: noup flag(s) set (OSDMAP_FLAGS) 2026-03-10T11:37:44.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:44 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd set", "key": "noup"}]': finished 2026-03-10T11:37:44.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:44 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["1", "7", "2"]}]: dispatch 2026-03-10T11:37:44.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:44 vm09 ceph-mon[54793]: osdmap e308: 8 total, 8 up, 8 in 2026-03-10T11:37:44.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:44 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd down", "ids": ["1", "7", "2"]}]: dispatch 2026-03-10T11:37:45.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:45 vm09 ceph-mon[54793]: Health check failed: 3 osds down (OSD_DOWN) 2026-03-10T11:37:45.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:45 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd down", "ids": ["1", "7", "2"]}]': finished 2026-03-10T11:37:45.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:45 vm09 ceph-mon[54793]: osdmap e309: 8 total, 5 up, 8 in 2026-03-10T11:37:45.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:45 vm09 ceph-mon[54793]: osd.7 marked itself dead as of e309 2026-03-10T11:37:45.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:45 vm06 ceph-mon[49228]: Health check failed: 3 osds down (OSD_DOWN) 2026-03-10T11:37:45.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:45 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd down", "ids": ["1", "7", "2"]}]': finished 2026-03-10T11:37:45.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:45 vm06 ceph-mon[49228]: osdmap e309: 8 total, 5 up, 8 in 2026-03-10T11:37:45.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:45 vm06 ceph-mon[49228]: osd.7 marked itself dead as of e309 2026-03-10T11:37:45.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:45 vm06 ceph-mon[57405]: Health check failed: 3 osds down (OSD_DOWN) 2026-03-10T11:37:45.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:45 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd down", "ids": ["1", "7", "2"]}]': finished 2026-03-10T11:37:45.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:45 vm06 ceph-mon[57405]: osdmap e309: 8 total, 5 up, 8 in 2026-03-10T11:37:45.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:45 vm06 ceph-mon[57405]: osd.7 marked itself dead as of e309 2026-03-10T11:37:45.980 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:37:45 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:37:45.666+0000 7f5097e22640 -1 osd.7 310 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:46.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:46 vm09 ceph-mon[54793]: Monitor daemon marked osd.7 down, but it is still running 2026-03-10T11:37:46.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:46 vm09 ceph-mon[54793]: map e309 wrongly marked me down at e309 2026-03-10T11:37:46.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:46 vm09 ceph-mon[54793]: pgmap v415: 196 pgs: 51 stale+active+clean, 32 unknown, 113 active+clean; 455 KiB data, 498 MiB used, 159 GiB / 160 GiB avail; 761 B/s rd, 0 op/s; 81 B/s, 2 objects/s recovering 2026-03-10T11:37:46.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:46 vm09 ceph-mon[54793]: osdmap e310: 8 total, 5 up, 8 in 2026-03-10T11:37:46.480 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:37:46 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:37:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:46 vm06 ceph-mon[49228]: Monitor daemon marked osd.7 down, but it is still running 2026-03-10T11:37:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:46 vm06 ceph-mon[49228]: map e309 wrongly marked me down at e309 2026-03-10T11:37:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:46 vm06 ceph-mon[49228]: pgmap v415: 196 pgs: 51 stale+active+clean, 32 unknown, 113 active+clean; 455 KiB data, 498 MiB used, 159 GiB / 160 GiB avail; 761 B/s rd, 0 op/s; 81 B/s, 2 objects/s recovering 2026-03-10T11:37:46.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:46 vm06 ceph-mon[49228]: osdmap e310: 8 total, 5 up, 8 in 2026-03-10T11:37:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:46 vm06 ceph-mon[57405]: Monitor daemon marked osd.7 down, but it is still running 2026-03-10T11:37:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:46 vm06 ceph-mon[57405]: map e309 wrongly marked me down at e309 2026-03-10T11:37:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:46 vm06 ceph-mon[57405]: pgmap v415: 196 pgs: 51 stale+active+clean, 32 unknown, 113 active+clean; 455 KiB data, 498 MiB used, 159 GiB / 160 GiB avail; 761 B/s rd, 0 op/s; 81 B/s, 2 objects/s recovering 2026-03-10T11:37:46.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:46 vm06 ceph-mon[57405]: osdmap e310: 8 total, 5 up, 8 in 2026-03-10T11:37:47.031 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:37:46 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1[66565]: 2026-03-10T11:37:46.693+0000 7f2f5307b640 -1 osd.1 310 osdmap NOUP flag is set, waiting for it to clear 2026-03-10T11:37:47.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:47 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:47.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:47 vm09 ceph-mon[54793]: Monitor daemon marked osd.1 down, but it is still running 2026-03-10T11:37:47.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:47 vm09 ceph-mon[54793]: map e310 wrongly marked me down at e309 2026-03-10T11:37:47.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:47 vm09 ceph-mon[54793]: osd.1 marked itself dead as of e310 2026-03-10T11:37:47.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:47 vm09 ceph-mon[54793]: osd.2 marked itself dead as of e310 2026-03-10T11:37:47.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:47 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:47.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:47 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:47.480 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:37:47 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:37:47.173+0000 7f5092c38640 -1 osd.7 311 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:37:47.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[49228]: Monitor daemon marked osd.1 down, but it is still running 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[49228]: map e310 wrongly marked me down at e309 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[49228]: osd.1 marked itself dead as of e310 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[49228]: osd.2 marked itself dead as of e310 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[57405]: Monitor daemon marked osd.1 down, but it is still running 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[57405]: map e310 wrongly marked me down at e309 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[57405]: osd.1 marked itself dead as of e310 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[57405]: osd.2 marked itself dead as of e310 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:47.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:47 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:47.531 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:37:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1[66565]: 2026-03-10T11:37:47.183+0000 7f2f4de91640 -1 osd.1 311 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:37:48.199 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:37:47 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2[71586]: 2026-03-10T11:37:47.912+0000 7f4b52556640 -1 osd.2 311 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T11:37:48.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:48 vm09 ceph-mon[54793]: Monitor daemon marked osd.2 down, but it is still running 2026-03-10T11:37:48.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:48 vm09 ceph-mon[54793]: map e310 wrongly marked me down at e309 2026-03-10T11:37:48.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:48 vm09 ceph-mon[54793]: pgmap v417: 196 pgs: 25 active+undersized, 15 active+undersized+degraded+wait, 11 undersized+degraded+peered+wait, 1 stale+active+clean, 1 unknown, 50 active+undersized+wait, 32 undersized+peered+wait, 14 active+undersized+degraded, 47 active+clean; 455 KiB data, 503 MiB used, 159 GiB / 160 GiB avail; 190/597 objects degraded (31.826%) 2026-03-10T11:37:48.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:48 vm09 ceph-mon[54793]: Health check failed: Degraded data redundancy: 190/597 objects degraded (31.826%), 40 pgs degraded (PG_DEGRADED) 2026-03-10T11:37:48.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:48 vm09 ceph-mon[54793]: Health check cleared: OSDMAP_FLAGS (was: noup flag(s) set) 2026-03-10T11:37:48.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:48 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:48.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:48 vm09 ceph-mon[54793]: osdmap e311: 8 total, 5 up, 8 in 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[49228]: Monitor daemon marked osd.2 down, but it is still running 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[49228]: map e310 wrongly marked me down at e309 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[49228]: pgmap v417: 196 pgs: 25 active+undersized, 15 active+undersized+degraded+wait, 11 undersized+degraded+peered+wait, 1 stale+active+clean, 1 unknown, 50 active+undersized+wait, 32 undersized+peered+wait, 14 active+undersized+degraded, 47 active+clean; 455 KiB data, 503 MiB used, 159 GiB / 160 GiB avail; 190/597 objects degraded (31.826%) 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[49228]: Health check failed: Degraded data redundancy: 190/597 objects degraded (31.826%), 40 pgs degraded (PG_DEGRADED) 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[49228]: Health check cleared: OSDMAP_FLAGS (was: noup flag(s) set) 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[49228]: osdmap e311: 8 total, 5 up, 8 in 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[57405]: Monitor daemon marked osd.2 down, but it is still running 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[57405]: map e310 wrongly marked me down at e309 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[57405]: pgmap v417: 196 pgs: 25 active+undersized, 15 active+undersized+degraded+wait, 11 undersized+degraded+peered+wait, 1 stale+active+clean, 1 unknown, 50 active+undersized+wait, 32 undersized+peered+wait, 14 active+undersized+degraded, 47 active+clean; 455 KiB data, 503 MiB used, 159 GiB / 160 GiB avail; 190/597 objects degraded (31.826%) 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[57405]: Health check failed: Degraded data redundancy: 190/597 objects degraded (31.826%), 40 pgs degraded (PG_DEGRADED) 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[57405]: Health check cleared: OSDMAP_FLAGS (was: noup flag(s) set) 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:48 vm06 ceph-mon[57405]: osdmap e311: 8 total, 5 up, 8 in 2026-03-10T11:37:49.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:49 vm09 ceph-mon[54793]: Health check cleared: OSD_DOWN (was: 3 osds down) 2026-03-10T11:37:49.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:49 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:37:49.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:49 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:37:49.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:49 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:37:49.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:49 vm09 ceph-mon[54793]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:37:49.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:49 vm09 ceph-mon[54793]: osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054] boot 2026-03-10T11:37:49.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:49 vm09 ceph-mon[54793]: osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931] boot 2026-03-10T11:37:49.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:49 vm09 ceph-mon[54793]: osdmap e312: 8 total, 8 up, 8 in 2026-03-10T11:37:49.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:49 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[49228]: Health check cleared: OSD_DOWN (was: 3 osds down) 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[49228]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[49228]: osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054] boot 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[49228]: osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931] boot 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[49228]: osdmap e312: 8 total, 8 up, 8 in 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[57405]: Health check cleared: OSD_DOWN (was: 3 osds down) 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[57405]: osd.7 [v2:192.168.123.109:6824/1888228353,v1:192.168.123.109:6825/1888228353] boot 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[57405]: osd.1 [v2:192.168.123.106:6810/193031054,v1:192.168.123.106:6811/193031054] boot 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[57405]: osd.2 [v2:192.168.123.106:6818/316237931,v1:192.168.123.106:6819/316237931] boot 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[57405]: osdmap e312: 8 total, 8 up, 8 in 2026-03-10T11:37:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:49 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:50.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:50 vm09 ceph-mon[54793]: pgmap v420: 196 pgs: 25 active+undersized, 15 active+undersized+degraded+wait, 11 undersized+degraded+peered+wait, 15 stale+active+clean, 1 unknown, 50 active+undersized+wait, 32 undersized+peered+wait, 14 active+undersized+degraded, 33 active+clean; 455 KiB data, 503 MiB used, 159 GiB / 160 GiB avail; 190/597 objects degraded (31.826%) 2026-03-10T11:37:50.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:50 vm09 ceph-mon[54793]: osdmap e313: 8 total, 8 up, 8 in 2026-03-10T11:37:50.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:50 vm06 ceph-mon[49228]: pgmap v420: 196 pgs: 25 active+undersized, 15 active+undersized+degraded+wait, 11 undersized+degraded+peered+wait, 15 stale+active+clean, 1 unknown, 50 active+undersized+wait, 32 undersized+peered+wait, 14 active+undersized+degraded, 33 active+clean; 455 KiB data, 503 MiB used, 159 GiB / 160 GiB avail; 190/597 objects degraded (31.826%) 2026-03-10T11:37:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:50 vm06 ceph-mon[49228]: osdmap e313: 8 total, 8 up, 8 in 2026-03-10T11:37:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:50 vm06 ceph-mon[57405]: pgmap v420: 196 pgs: 25 active+undersized, 15 active+undersized+degraded+wait, 11 undersized+degraded+peered+wait, 15 stale+active+clean, 1 unknown, 50 active+undersized+wait, 32 undersized+peered+wait, 14 active+undersized+degraded, 33 active+clean; 455 KiB data, 503 MiB used, 159 GiB / 160 GiB avail; 190/597 objects degraded (31.826%) 2026-03-10T11:37:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:50 vm06 ceph-mon[57405]: osdmap e313: 8 total, 8 up, 8 in 2026-03-10T11:37:51.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:37:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:37:50] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:37:51.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:51 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:51.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:51 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:51 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:51.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:51 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:51 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3698709347' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:51.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:51 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:52.238 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_read_wait_for_complete_and_cb_error PASSED [ 79%] 2026-03-10T11:37:52.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:52 vm09 ceph-mon[54793]: pgmap v422: 196 pgs: 196 active+clean; 455 KiB data, 507 MiB used, 159 GiB / 160 GiB avail; 519 B/s rd, 0 op/s 2026-03-10T11:37:52.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:52 vm09 ceph-mon[54793]: Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 190/597 objects degraded (31.826%), 40 pgs degraded) 2026-03-10T11:37:52.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:52 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:52.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:52 vm09 ceph-mon[54793]: osdmap e314: 8 total, 8 up, 8 in 2026-03-10T11:37:52.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:52 vm06 ceph-mon[49228]: pgmap v422: 196 pgs: 196 active+clean; 455 KiB data, 507 MiB used, 159 GiB / 160 GiB avail; 519 B/s rd, 0 op/s 2026-03-10T11:37:52.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:52 vm06 ceph-mon[49228]: Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 190/597 objects degraded (31.826%), 40 pgs degraded) 2026-03-10T11:37:52.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:52 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:52.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:52 vm06 ceph-mon[49228]: osdmap e314: 8 total, 8 up, 8 in 2026-03-10T11:37:52.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:52 vm06 ceph-mon[57405]: pgmap v422: 196 pgs: 196 active+clean; 455 KiB data, 507 MiB used, 159 GiB / 160 GiB avail; 519 B/s rd, 0 op/s 2026-03-10T11:37:52.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:52 vm06 ceph-mon[57405]: Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 190/597 objects degraded (31.826%), 40 pgs degraded) 2026-03-10T11:37:52.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:52 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:52.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:52 vm06 ceph-mon[57405]: osdmap e314: 8 total, 8 up, 8 in 2026-03-10T11:37:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:53 vm09 ceph-mon[54793]: osdmap e315: 8 total, 8 up, 8 in 2026-03-10T11:37:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:53 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:37:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:53 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:37:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:53 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:37:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:53 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:53.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:53 vm06 ceph-mon[49228]: osdmap e315: 8 total, 8 up, 8 in 2026-03-10T11:37:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:53 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:37:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:53 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:37:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:53 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:37:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:53 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:53 vm06 ceph-mon[57405]: osdmap e315: 8 total, 8 up, 8 in 2026-03-10T11:37:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:53 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:37:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:53 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:37:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:53 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:37:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:53 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:54.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:54 vm06 ceph-mon[49228]: pgmap v425: 164 pgs: 164 active+clean; 455 KiB data, 507 MiB used, 159 GiB / 160 GiB avail; 1.4 KiB/s rd, 1 op/s 2026-03-10T11:37:54.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:54 vm06 ceph-mon[49228]: osdmap e316: 8 total, 8 up, 8 in 2026-03-10T11:37:54.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:54 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:54.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:54 vm06 ceph-mon[57405]: pgmap v425: 164 pgs: 164 active+clean; 455 KiB data, 507 MiB used, 159 GiB / 160 GiB avail; 1.4 KiB/s rd, 1 op/s 2026-03-10T11:37:54.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:54 vm06 ceph-mon[57405]: osdmap e316: 8 total, 8 up, 8 in 2026-03-10T11:37:54.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:54 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:54.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:54 vm09 ceph-mon[54793]: pgmap v425: 164 pgs: 164 active+clean; 455 KiB data, 507 MiB used, 159 GiB / 160 GiB avail; 1.4 KiB/s rd, 1 op/s 2026-03-10T11:37:54.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:54 vm09 ceph-mon[54793]: osdmap e316: 8 total, 8 up, 8 in 2026-03-10T11:37:54.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:54 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:37:55.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:55 vm06 ceph-mon[49228]: osdmap e317: 8 total, 8 up, 8 in 2026-03-10T11:37:55.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:55 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1037198685' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:55.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:55 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:55.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:55 vm06 ceph-mon[57405]: osdmap e317: 8 total, 8 up, 8 in 2026-03-10T11:37:55.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:55 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1037198685' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:55.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:55 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:55.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:55 vm09 ceph-mon[54793]: osdmap e317: 8 total, 8 up, 8 in 2026-03-10T11:37:55.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:55 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1037198685' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:55.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:55 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:56.259 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_lock PASSED [ 80%] 2026-03-10T11:37:56.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:56 vm06 ceph-mon[49228]: pgmap v428: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 507 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:56.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:56 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:56.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:56 vm06 ceph-mon[49228]: osdmap e318: 8 total, 8 up, 8 in 2026-03-10T11:37:56.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:56 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:56.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:56 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:56.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:56 vm06 ceph-mon[57405]: pgmap v428: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 507 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:56.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:56 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:56.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:56 vm06 ceph-mon[57405]: osdmap e318: 8 total, 8 up, 8 in 2026-03-10T11:37:56.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:56 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:56.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:56 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:56 vm09 ceph-mon[54793]: pgmap v428: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 507 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:37:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:56 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:37:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:56 vm09 ceph-mon[54793]: osdmap e318: 8 total, 8 up, 8 in 2026-03-10T11:37:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:56 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:37:56.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:56 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:37:56.730 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:37:56 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:37:57.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:57 vm06 ceph-mon[49228]: osdmap e319: 8 total, 8 up, 8 in 2026-03-10T11:37:57.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:57 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:57.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:57 vm06 ceph-mon[57405]: osdmap e319: 8 total, 8 up, 8 in 2026-03-10T11:37:57.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:57 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:57.729 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:57 vm09 ceph-mon[54793]: osdmap e319: 8 total, 8 up, 8 in 2026-03-10T11:37:57.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:57 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:37:58.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:58 vm09 ceph-mon[54793]: pgmap v431: 164 pgs: 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:58.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:58 vm09 ceph-mon[54793]: osdmap e320: 8 total, 8 up, 8 in 2026-03-10T11:37:58.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:58 vm06 ceph-mon[49228]: pgmap v431: 164 pgs: 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:58.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:58 vm06 ceph-mon[49228]: osdmap e320: 8 total, 8 up, 8 in 2026-03-10T11:37:58.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:58 vm06 ceph-mon[57405]: pgmap v431: 164 pgs: 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:37:58.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:58 vm06 ceph-mon[57405]: osdmap e320: 8 total, 8 up, 8 in 2026-03-10T11:37:59.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:59 vm09 ceph-mon[54793]: osdmap e321: 8 total, 8 up, 8 in 2026-03-10T11:37:59.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:59 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/4163054735' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:59.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:37:59 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:59.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:59 vm06 ceph-mon[49228]: osdmap e321: 8 total, 8 up, 8 in 2026-03-10T11:37:59.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:59 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/4163054735' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:59.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:37:59 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:59 vm06 ceph-mon[57405]: osdmap e321: 8 total, 8 up, 8 in 2026-03-10T11:37:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:59 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/4163054735' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:37:59.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:37:59 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:00.443 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_execute PASSED [ 81%] 2026-03-10T11:38:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:00 vm09 ceph-mon[54793]: pgmap v434: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:00 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:00 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:00 vm09 ceph-mon[54793]: osdmap e322: 8 total, 8 up, 8 in 2026-03-10T11:38:00.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:00 vm06 ceph-mon[49228]: pgmap v434: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:00.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:00 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:00.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:00 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:00.742 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:00 vm06 ceph-mon[49228]: osdmap e322: 8 total, 8 up, 8 in 2026-03-10T11:38:00.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:00 vm06 ceph-mon[57405]: pgmap v434: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:00.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:00 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:00.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:00 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:00.742 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:00 vm06 ceph-mon[57405]: osdmap e322: 8 total, 8 up, 8 in 2026-03-10T11:38:01.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:38:00 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:38:00] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:38:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:01 vm09 ceph-mon[54793]: osdmap e323: 8 total, 8 up, 8 in 2026-03-10T11:38:01.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:01 vm09 ceph-mon[54793]: pgmap v437: 164 pgs: 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:01.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:01 vm06 ceph-mon[49228]: osdmap e323: 8 total, 8 up, 8 in 2026-03-10T11:38:01.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:01 vm06 ceph-mon[49228]: pgmap v437: 164 pgs: 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:01.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:01 vm06 ceph-mon[57405]: osdmap e323: 8 total, 8 up, 8 in 2026-03-10T11:38:01.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:01 vm06 ceph-mon[57405]: pgmap v437: 164 pgs: 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:02.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:02 vm06 ceph-mon[49228]: osdmap e324: 8 total, 8 up, 8 in 2026-03-10T11:38:02.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:02 vm06 ceph-mon[57405]: osdmap e324: 8 total, 8 up, 8 in 2026-03-10T11:38:02.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:02 vm09 ceph-mon[54793]: osdmap e324: 8 total, 8 up, 8 in 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[49228]: osdmap e325: 8 total, 8 up, 8 in 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/4041971142' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[49228]: pgmap v440: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[49228]: osdmap e326: 8 total, 8 up, 8 in 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[57405]: osdmap e325: 8 total, 8 up, 8 in 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/4041971142' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[57405]: pgmap v440: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:03.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:03 vm06 ceph-mon[57405]: osdmap e326: 8 total, 8 up, 8 in 2026-03-10T11:38:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:03 vm09 ceph-mon[54793]: osdmap e325: 8 total, 8 up, 8 in 2026-03-10T11:38:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:03 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/4041971142' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:03 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:03 vm09 ceph-mon[54793]: pgmap v440: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:03 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:03.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:03 vm09 ceph-mon[54793]: osdmap e326: 8 total, 8 up, 8 in 2026-03-10T11:38:04.506 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_execute PASSED [ 82%] 2026-03-10T11:38:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:05 vm06 ceph-mon[49228]: osdmap e327: 8 total, 8 up, 8 in 2026-03-10T11:38:05.781 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:05 vm06 ceph-mon[49228]: pgmap v443: 164 pgs: 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:05 vm06 ceph-mon[57405]: osdmap e327: 8 total, 8 up, 8 in 2026-03-10T11:38:05.781 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:05 vm06 ceph-mon[57405]: pgmap v443: 164 pgs: 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:05 vm09 ceph-mon[54793]: osdmap e327: 8 total, 8 up, 8 in 2026-03-10T11:38:05.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:05 vm09 ceph-mon[54793]: pgmap v443: 164 pgs: 164 active+clean; 455 KiB data, 508 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:06.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:06 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:06.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:06 vm09 ceph-mon[54793]: osdmap e328: 8 total, 8 up, 8 in 2026-03-10T11:38:06.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:06 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:06.730 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:38:06 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:38:07.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:06 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:07.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:06 vm06 ceph-mon[49228]: osdmap e328: 8 total, 8 up, 8 in 2026-03-10T11:38:07.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:06 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:07.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:06 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:07.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:06 vm06 ceph-mon[57405]: osdmap e328: 8 total, 8 up, 8 in 2026-03-10T11:38:07.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:06 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:07 vm09 ceph-mon[54793]: osdmap e329: 8 total, 8 up, 8 in 2026-03-10T11:38:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:07 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1155572520' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:07 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:07.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:07 vm09 ceph-mon[54793]: pgmap v446: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 513 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:08.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:07 vm06 ceph-mon[49228]: osdmap e329: 8 total, 8 up, 8 in 2026-03-10T11:38:08.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:07 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1155572520' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:08.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:07 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:08.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:07 vm06 ceph-mon[49228]: pgmap v446: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 513 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:08.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:07 vm06 ceph-mon[57405]: osdmap e329: 8 total, 8 up, 8 in 2026-03-10T11:38:08.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:07 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1155572520' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:08.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:07 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:08.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:07 vm06 ceph-mon[57405]: pgmap v446: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 513 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:08.591 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_aio_setxattr PASSED [ 83%] 2026-03-10T11:38:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:08 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:08.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:08 vm09 ceph-mon[54793]: osdmap e330: 8 total, 8 up, 8 in 2026-03-10T11:38:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:08 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:09.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:08 vm06 ceph-mon[49228]: osdmap e330: 8 total, 8 up, 8 in 2026-03-10T11:38:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:08 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:09.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:08 vm06 ceph-mon[57405]: osdmap e330: 8 total, 8 up, 8 in 2026-03-10T11:38:09.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:09 vm09 ceph-mon[54793]: osdmap e331: 8 total, 8 up, 8 in 2026-03-10T11:38:09.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:09 vm09 ceph-mon[54793]: pgmap v449: 164 pgs: 164 active+clean; 455 KiB data, 513 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:09 vm06 ceph-mon[49228]: osdmap e331: 8 total, 8 up, 8 in 2026-03-10T11:38:10.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:09 vm06 ceph-mon[49228]: pgmap v449: 164 pgs: 164 active+clean; 455 KiB data, 513 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:09 vm06 ceph-mon[57405]: osdmap e331: 8 total, 8 up, 8 in 2026-03-10T11:38:10.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:09 vm06 ceph-mon[57405]: pgmap v449: 164 pgs: 164 active+clean; 455 KiB data, 513 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:10.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:10 vm09 ceph-mon[54793]: osdmap e332: 8 total, 8 up, 8 in 2026-03-10T11:38:10.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:10 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:38:10.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:10 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app1"}]: dispatch 2026-03-10T11:38:10.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:10 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app1"}]: dispatch 2026-03-10T11:38:10.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:10 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:11.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:38:10 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:38:10] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:38:11.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:10 vm06 ceph-mon[49228]: osdmap e332: 8 total, 8 up, 8 in 2026-03-10T11:38:11.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:10 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:38:11.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:10 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app1"}]: dispatch 2026-03-10T11:38:11.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:10 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app1"}]: dispatch 2026-03-10T11:38:11.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:10 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:10 vm06 ceph-mon[57405]: osdmap e332: 8 total, 8 up, 8 in 2026-03-10T11:38:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:10 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T11:38:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:10 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app1"}]: dispatch 2026-03-10T11:38:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:10 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app1"}]: dispatch 2026-03-10T11:38:11.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:10 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:11 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "test_pool","app": "app1"}]': finished 2026-03-10T11:38:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:11 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2"}]: dispatch 2026-03-10T11:38:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:11 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2","yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:38:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:11 vm09 ceph-mon[54793]: osdmap e333: 8 total, 8 up, 8 in 2026-03-10T11:38:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:11 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2","yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:38:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:11 vm09 ceph-mon[54793]: pgmap v452: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:11 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2","yes_i_really_mean_it": true}]': finished 2026-03-10T11:38:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:11 vm09 ceph-mon[54793]: osdmap e334: 8 total, 8 up, 8 in 2026-03-10T11:38:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:11 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"dne","key":"key","value":"key"}]: dispatch 2026-03-10T11:38:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:11 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:11.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:11 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "test_pool","app": "app1"}]': finished 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2"}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2","yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[49228]: osdmap e333: 8 total, 8 up, 8 in 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2","yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[49228]: pgmap v452: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2","yes_i_really_mean_it": true}]': finished 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[49228]: osdmap e334: 8 total, 8 up, 8 in 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"dne","key":"key","value":"key"}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "test_pool","app": "app1"}]': finished 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2"}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2","yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[57405]: osdmap e333: 8 total, 8 up, 8 in 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2","yes_i_really_mean_it": true}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[57405]: pgmap v452: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "test_pool","app": "app2","yes_i_really_mean_it": true}]': finished 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[57405]: osdmap e334: 8 total, 8 up, 8 in 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"dne","key":"key","value":"key"}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:12.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:11 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:13 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key1","value":"val1"}]': finished 2026-03-10T11:38:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:13 vm09 ceph-mon[54793]: osdmap e335: 8 total, 8 up, 8 in 2026-03-10T11:38:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:13 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key2","value":"val2"}]: dispatch 2026-03-10T11:38:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:13 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key2","value":"val2"}]: dispatch 2026-03-10T11:38:13.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:13 vm09 ceph-mon[54793]: pgmap v455: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:14.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:13 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key1","value":"val1"}]': finished 2026-03-10T11:38:14.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:13 vm06 ceph-mon[49228]: osdmap e335: 8 total, 8 up, 8 in 2026-03-10T11:38:14.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:13 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key2","value":"val2"}]: dispatch 2026-03-10T11:38:14.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:13 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key2","value":"val2"}]: dispatch 2026-03-10T11:38:14.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:13 vm06 ceph-mon[49228]: pgmap v455: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:14.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:13 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key1","value":"val1"}]': finished 2026-03-10T11:38:14.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:13 vm06 ceph-mon[57405]: osdmap e335: 8 total, 8 up, 8 in 2026-03-10T11:38:14.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:13 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key2","value":"val2"}]: dispatch 2026-03-10T11:38:14.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:13 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key2","value":"val2"}]: dispatch 2026-03-10T11:38:14.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:13 vm06 ceph-mon[57405]: pgmap v455: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:14.932 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:14 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key2","value":"val2"}]': finished 2026-03-10T11:38:14.932 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:14 vm06 ceph-mon[49228]: osdmap e336: 8 total, 8 up, 8 in 2026-03-10T11:38:14.932 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:14 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app2","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:14.932 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:14 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app2","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:14.933 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:14 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key2","value":"val2"}]': finished 2026-03-10T11:38:14.933 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:14 vm06 ceph-mon[57405]: osdmap e336: 8 total, 8 up, 8 in 2026-03-10T11:38:14.933 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:14 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app2","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:14.933 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:14 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app2","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:14.979 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:14 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application set","pool":"test_pool","app":"app1","key":"key2","value":"val2"}]': finished 2026-03-10T11:38:14.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:14 vm09 ceph-mon[54793]: osdmap e336: 8 total, 8 up, 8 in 2026-03-10T11:38:14.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:14 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app2","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:14.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:14 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application set","pool":"test_pool","app":"app2","key":"key1","value":"val1"}]: dispatch 2026-03-10T11:38:15.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:15 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application set","pool":"test_pool","app":"app2","key":"key1","value":"val1"}]': finished 2026-03-10T11:38:15.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:15 vm09 ceph-mon[54793]: osdmap e337: 8 total, 8 up, 8 in 2026-03-10T11:38:15.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:15 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application rm","pool":"test_pool","app":"app1","key":"key1"}]: dispatch 2026-03-10T11:38:15.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:15 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application rm","pool":"test_pool","app":"app1","key":"key1"}]: dispatch 2026-03-10T11:38:15.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:15 vm09 ceph-mon[54793]: pgmap v458: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:16.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:15 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application set","pool":"test_pool","app":"app2","key":"key1","value":"val1"}]': finished 2026-03-10T11:38:16.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:15 vm06 ceph-mon[49228]: osdmap e337: 8 total, 8 up, 8 in 2026-03-10T11:38:16.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:15 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application rm","pool":"test_pool","app":"app1","key":"key1"}]: dispatch 2026-03-10T11:38:16.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:15 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application rm","pool":"test_pool","app":"app1","key":"key1"}]: dispatch 2026-03-10T11:38:16.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:15 vm06 ceph-mon[49228]: pgmap v458: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:16.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:15 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application set","pool":"test_pool","app":"app2","key":"key1","value":"val1"}]': finished 2026-03-10T11:38:16.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:15 vm06 ceph-mon[57405]: osdmap e337: 8 total, 8 up, 8 in 2026-03-10T11:38:16.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:15 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix":"osd pool application rm","pool":"test_pool","app":"app1","key":"key1"}]: dispatch 2026-03-10T11:38:16.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:15 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix":"osd pool application rm","pool":"test_pool","app":"app1","key":"key1"}]: dispatch 2026-03-10T11:38:16.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:15 vm06 ceph-mon[57405]: pgmap v458: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:16.654 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:38:16 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:38:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:16 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application rm","pool":"test_pool","app":"app1","key":"key1"}]': finished 2026-03-10T11:38:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:16 vm09 ceph-mon[54793]: osdmap e338: 8 total, 8 up, 8 in 2026-03-10T11:38:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:16 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:16 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:16 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:16 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:16.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:16 vm09 ceph-mon[54793]: osdmap e339: 8 total, 8 up, 8 in 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application rm","pool":"test_pool","app":"app1","key":"key1"}]': finished 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[49228]: osdmap e338: 8 total, 8 up, 8 in 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[49228]: osdmap e339: 8 total, 8 up, 8 in 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix":"osd pool application rm","pool":"test_pool","app":"app1","key":"key1"}]': finished 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[57405]: osdmap e338: 8 total, 8 up, 8 in 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3228401783' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:17.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:16 vm06 ceph-mon[57405]: osdmap e339: 8 total, 8 up, 8 in 2026-03-10T11:38:17.654 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_applications PASSED [ 84%] 2026-03-10T11:38:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:17 vm09 ceph-mon[54793]: pgmap v461: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:17.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:17 vm09 ceph-mon[54793]: osdmap e340: 8 total, 8 up, 8 in 2026-03-10T11:38:18.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:17 vm06 ceph-mon[49228]: pgmap v461: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:18.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:17 vm06 ceph-mon[49228]: osdmap e340: 8 total, 8 up, 8 in 2026-03-10T11:38:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:17 vm06 ceph-mon[57405]: pgmap v461: 196 pgs: 196 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:18.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:17 vm06 ceph-mon[57405]: osdmap e340: 8 total, 8 up, 8 in 2026-03-10T11:38:19.688 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:19 vm06 ceph-mon[49228]: osdmap e341: 8 total, 8 up, 8 in 2026-03-10T11:38:19.689 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:19 vm06 ceph-mon[57405]: osdmap e341: 8 total, 8 up, 8 in 2026-03-10T11:38:19.689 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:19 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1031667246' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:19.693 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:19 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:19.693 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:19 vm06 ceph-mon[57405]: pgmap v464: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:19.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:19 vm09 ceph-mon[54793]: osdmap e341: 8 total, 8 up, 8 in 2026-03-10T11:38:19.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:19 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1031667246' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:19.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:19 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:19.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:19 vm09 ceph-mon[54793]: pgmap v464: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:20.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:19 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1031667246' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:20.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:19 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:20.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:19 vm06 ceph-mon[49228]: pgmap v464: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 518 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:20.680 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_service_daemon PASSED [ 85%] 2026-03-10T11:38:20.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:20 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:20.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:20 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:20.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:20 vm09 ceph-mon[54793]: osdmap e342: 8 total, 8 up, 8 in 2026-03-10T11:38:20.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:20 vm09 ceph-mon[54793]: osdmap e343: 8 total, 8 up, 8 in 2026-03-10T11:38:21.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:38:20 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:38:20] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:38:21.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:20 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:21.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:20 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:21.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:20 vm06 ceph-mon[49228]: osdmap e342: 8 total, 8 up, 8 in 2026-03-10T11:38:21.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:20 vm06 ceph-mon[49228]: osdmap e343: 8 total, 8 up, 8 in 2026-03-10T11:38:21.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:20 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:21.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:20 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:21.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:20 vm06 ceph-mon[57405]: osdmap e342: 8 total, 8 up, 8 in 2026-03-10T11:38:21.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:20 vm06 ceph-mon[57405]: osdmap e343: 8 total, 8 up, 8 in 2026-03-10T11:38:21.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:21 vm09 ceph-mon[54793]: pgmap v467: 164 pgs: 164 active+clean; 455 KiB data, 523 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:22.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:21 vm06 ceph-mon[49228]: pgmap v467: 164 pgs: 164 active+clean; 455 KiB data, 523 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:22.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:21 vm06 ceph-mon[57405]: pgmap v467: 164 pgs: 164 active+clean; 455 KiB data, 523 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:22.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:22 vm09 ceph-mon[54793]: osdmap e344: 8 total, 8 up, 8 in 2026-03-10T11:38:22.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:22 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/927168997' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:22.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:22 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:23.030 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:22 vm06 ceph-mon[49228]: osdmap e344: 8 total, 8 up, 8 in 2026-03-10T11:38:23.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:22 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/927168997' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:23.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:22 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:23.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:22 vm06 ceph-mon[57405]: osdmap e344: 8 total, 8 up, 8 in 2026-03-10T11:38:23.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:22 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/927168997' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:23.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:22 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:23.717 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx::test_alignment PASSED [ 86%] 2026-03-10T11:38:23.735 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:23 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:23.735 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:23 vm09 ceph-mon[54793]: osdmap e345: 8 total, 8 up, 8 in 2026-03-10T11:38:23.739 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:23 vm09 ceph-mon[54793]: pgmap v470: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 523 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:23.739 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:23 vm09 ceph-mon[54793]: osdmap e346: 8 total, 8 up, 8 in 2026-03-10T11:38:23.739 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:23 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2109172064' entity='client.admin' cmd=[{"prefix": "osd erasure-code-profile set", "name": "testprofile-test-ec", "profile": ["k=2", "m=1", "crush-failure-domain=osd"]}]: dispatch 2026-03-10T11:38:23.739 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:23 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd erasure-code-profile set", "name": "testprofile-test-ec", "profile": ["k=2", "m=1", "crush-failure-domain=osd"]}]: dispatch 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[49228]: osdmap e345: 8 total, 8 up, 8 in 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[49228]: pgmap v470: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 523 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[49228]: osdmap e346: 8 total, 8 up, 8 in 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2109172064' entity='client.admin' cmd=[{"prefix": "osd erasure-code-profile set", "name": "testprofile-test-ec", "profile": ["k=2", "m=1", "crush-failure-domain=osd"]}]: dispatch 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd erasure-code-profile set", "name": "testprofile-test-ec", "profile": ["k=2", "m=1", "crush-failure-domain=osd"]}]: dispatch 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[57405]: osdmap e345: 8 total, 8 up, 8 in 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[57405]: pgmap v470: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 523 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[57405]: osdmap e346: 8 total, 8 up, 8 in 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2109172064' entity='client.admin' cmd=[{"prefix": "osd erasure-code-profile set", "name": "testprofile-test-ec", "profile": ["k=2", "m=1", "crush-failure-domain=osd"]}]: dispatch 2026-03-10T11:38:24.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:23 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd erasure-code-profile set", "name": "testprofile-test-ec", "profile": ["k=2", "m=1", "crush-failure-domain=osd"]}]: dispatch 2026-03-10T11:38:25.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:25 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd erasure-code-profile set", "name": "testprofile-test-ec", "profile": ["k=2", "m=1", "crush-failure-domain=osd"]}]': finished 2026-03-10T11:38:25.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:25 vm09 ceph-mon[54793]: osdmap e347: 8 total, 8 up, 8 in 2026-03-10T11:38:25.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:25 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2109172064' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 8, "pgp_num": 8, "pool": "test-ec", "pool_type": "erasure", "erasure_code_profile": "testprofile-test-ec"}]: dispatch 2026-03-10T11:38:25.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:25 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 8, "pgp_num": 8, "pool": "test-ec", "pool_type": "erasure", "erasure_code_profile": "testprofile-test-ec"}]: dispatch 2026-03-10T11:38:25.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:25 vm09 ceph-mon[54793]: pgmap v473: 164 pgs: 164 active+clean; 455 KiB data, 523 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:25.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:25 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd erasure-code-profile set", "name": "testprofile-test-ec", "profile": ["k=2", "m=1", "crush-failure-domain=osd"]}]': finished 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[49228]: osdmap e347: 8 total, 8 up, 8 in 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2109172064' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 8, "pgp_num": 8, "pool": "test-ec", "pool_type": "erasure", "erasure_code_profile": "testprofile-test-ec"}]: dispatch 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 8, "pgp_num": 8, "pool": "test-ec", "pool_type": "erasure", "erasure_code_profile": "testprofile-test-ec"}]: dispatch 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[49228]: pgmap v473: 164 pgs: 164 active+clean; 455 KiB data, 523 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd erasure-code-profile set", "name": "testprofile-test-ec", "profile": ["k=2", "m=1", "crush-failure-domain=osd"]}]': finished 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[57405]: osdmap e347: 8 total, 8 up, 8 in 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2109172064' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 8, "pgp_num": 8, "pool": "test-ec", "pool_type": "erasure", "erasure_code_profile": "testprofile-test-ec"}]: dispatch 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 8, "pgp_num": 8, "pool": "test-ec", "pool_type": "erasure", "erasure_code_profile": "testprofile-test-ec"}]: dispatch 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[57405]: pgmap v473: 164 pgs: 164 active+clean; 455 KiB data, 523 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:26.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:25 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:26.715 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:38:26 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:38:26.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:26 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:26.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:26 vm09 ceph-mon[54793]: osdmap e348: 8 total, 8 up, 8 in 2026-03-10T11:38:26.980 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:26 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:26 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:26 vm06 ceph-mon[49228]: osdmap e348: 8 total, 8 up, 8 in 2026-03-10T11:38:27.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:26 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:26 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:26 vm06 ceph-mon[57405]: osdmap e348: 8 total, 8 up, 8 in 2026-03-10T11:38:27.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:26 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:27 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pg_num": 8, "pgp_num": 8, "pool": "test-ec", "pool_type": "erasure", "erasure_code_profile": "testprofile-test-ec"}]': finished 2026-03-10T11:38:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:27 vm06 ceph-mon[49228]: osdmap e349: 8 total, 8 up, 8 in 2026-03-10T11:38:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:27 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/2109172064' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:27 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:28.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:27 vm06 ceph-mon[49228]: pgmap v476: 172 pgs: 8 unknown, 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:27 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pg_num": 8, "pgp_num": 8, "pool": "test-ec", "pool_type": "erasure", "erasure_code_profile": "testprofile-test-ec"}]': finished 2026-03-10T11:38:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:27 vm06 ceph-mon[57405]: osdmap e349: 8 total, 8 up, 8 in 2026-03-10T11:38:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:27 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/2109172064' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:27 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:28.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:27 vm06 ceph-mon[57405]: pgmap v476: 172 pgs: 8 unknown, 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:27 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pg_num": 8, "pgp_num": 8, "pool": "test-ec", "pool_type": "erasure", "erasure_code_profile": "testprofile-test-ec"}]': finished 2026-03-10T11:38:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:27 vm09 ceph-mon[54793]: osdmap e349: 8 total, 8 up, 8 in 2026-03-10T11:38:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:27 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/2109172064' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:27 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:28.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:27 vm09 ceph-mon[54793]: pgmap v476: 172 pgs: 8 unknown, 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:28.740 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctxEc::test_alignment PASSED [ 87%] 2026-03-10T11:38:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:28 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:28 vm06 ceph-mon[49228]: osdmap e350: 8 total, 8 up, 8 in 2026-03-10T11:38:29.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:28 vm06 ceph-mon[49228]: osdmap e351: 8 total, 8 up, 8 in 2026-03-10T11:38:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:28 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:28 vm06 ceph-mon[57405]: osdmap e350: 8 total, 8 up, 8 in 2026-03-10T11:38:29.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:28 vm06 ceph-mon[57405]: osdmap e351: 8 total, 8 up, 8 in 2026-03-10T11:38:29.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:28 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:29.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:28 vm09 ceph-mon[54793]: osdmap e350: 8 total, 8 up, 8 in 2026-03-10T11:38:29.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:28 vm09 ceph-mon[54793]: osdmap e351: 8 total, 8 up, 8 in 2026-03-10T11:38:30.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:29 vm09 ceph-mon[54793]: pgmap v479: 164 pgs: 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:30.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:29 vm06 ceph-mon[49228]: pgmap v479: 164 pgs: 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:30.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:29 vm06 ceph-mon[57405]: pgmap v479: 164 pgs: 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:31.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:38:30 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:38:30] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:38:31.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:30 vm06 ceph-mon[49228]: osdmap e352: 8 total, 8 up, 8 in 2026-03-10T11:38:31.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:30 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3267311612' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:31.031 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:30 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:31.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:30 vm06 ceph-mon[57405]: osdmap e352: 8 total, 8 up, 8 in 2026-03-10T11:38:31.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:30 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3267311612' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:31.031 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:30 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:31.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:30 vm09 ceph-mon[54793]: osdmap e352: 8 total, 8 up, 8 in 2026-03-10T11:38:31.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:30 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3267311612' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:31.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:30 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:31.858 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx2::test_get_last_version PASSED [ 89%] 2026-03-10T11:38:32.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:31 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:32.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:31 vm09 ceph-mon[54793]: osdmap e353: 8 total, 8 up, 8 in 2026-03-10T11:38:32.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:31 vm09 ceph-mon[54793]: pgmap v482: 196 pgs: 196 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:32.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:31 vm09 ceph-mon[54793]: osdmap e354: 8 total, 8 up, 8 in 2026-03-10T11:38:32.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:31 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:32.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:31 vm06 ceph-mon[49228]: osdmap e353: 8 total, 8 up, 8 in 2026-03-10T11:38:32.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:31 vm06 ceph-mon[49228]: pgmap v482: 196 pgs: 196 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:32.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:31 vm06 ceph-mon[49228]: osdmap e354: 8 total, 8 up, 8 in 2026-03-10T11:38:32.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:31 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:32.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:31 vm06 ceph-mon[57405]: osdmap e353: 8 total, 8 up, 8 in 2026-03-10T11:38:32.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:31 vm06 ceph-mon[57405]: pgmap v482: 196 pgs: 196 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:32.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:31 vm06 ceph-mon[57405]: osdmap e354: 8 total, 8 up, 8 in 2026-03-10T11:38:33.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:32 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:33.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:32 vm09 ceph-mon[54793]: osdmap e355: 8 total, 8 up, 8 in 2026-03-10T11:38:33.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:32 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/3540980992' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:33.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:32 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:33.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:32 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:33.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:32 vm06 ceph-mon[49228]: osdmap e355: 8 total, 8 up, 8 in 2026-03-10T11:38:33.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:32 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/3540980992' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:33.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:32 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:33.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:32 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:33.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:32 vm06 ceph-mon[57405]: osdmap e355: 8 total, 8 up, 8 in 2026-03-10T11:38:33.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:32 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/3540980992' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:33.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:32 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:34.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:33 vm09 ceph-mon[54793]: pgmap v485: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:34.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:33 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:34.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:33 vm09 ceph-mon[54793]: osdmap e356: 8 total, 8 up, 8 in 2026-03-10T11:38:34.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:33 vm06 ceph-mon[49228]: pgmap v485: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:34.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:33 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:34.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:33 vm06 ceph-mon[49228]: osdmap e356: 8 total, 8 up, 8 in 2026-03-10T11:38:34.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:33 vm06 ceph-mon[57405]: pgmap v485: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:34.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:33 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:34.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:33 vm06 ceph-mon[57405]: osdmap e356: 8 total, 8 up, 8 in 2026-03-10T11:38:34.887 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoctx2::test_get_stats PASSED [ 90%] 2026-03-10T11:38:36.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:35 vm09 ceph-mon[54793]: osdmap e357: 8 total, 8 up, 8 in 2026-03-10T11:38:36.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:35 vm09 ceph-mon[54793]: pgmap v488: 164 pgs: 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:36.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:35 vm06 ceph-mon[49228]: osdmap e357: 8 total, 8 up, 8 in 2026-03-10T11:38:36.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:35 vm06 ceph-mon[49228]: pgmap v488: 164 pgs: 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:35 vm06 ceph-mon[57405]: osdmap e357: 8 total, 8 up, 8 in 2026-03-10T11:38:36.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:35 vm06 ceph-mon[57405]: pgmap v488: 164 pgs: 164 active+clean; 455 KiB data, 524 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:36.730 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:38:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:38:37.229 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:36 vm09 ceph-mon[54793]: osdmap e358: 8 total, 8 up, 8 in 2026-03-10T11:38:37.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:36 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:36 vm06 ceph-mon[49228]: osdmap e358: 8 total, 8 up, 8 in 2026-03-10T11:38:37.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:36 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:36 vm06 ceph-mon[57405]: osdmap e358: 8 total, 8 up, 8 in 2026-03-10T11:38:37.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:36 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:37.944 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestObject::test_read PASSED [ 91%] 2026-03-10T11:38:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:37 vm09 ceph-mon[54793]: osdmap e359: 8 total, 8 up, 8 in 2026-03-10T11:38:38.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:37 vm09 ceph-mon[54793]: pgmap v491: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 525 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:38.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:37 vm06 ceph-mon[49228]: osdmap e359: 8 total, 8 up, 8 in 2026-03-10T11:38:38.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:37 vm06 ceph-mon[49228]: pgmap v491: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 525 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:37 vm06 ceph-mon[57405]: osdmap e359: 8 total, 8 up, 8 in 2026-03-10T11:38:38.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:37 vm06 ceph-mon[57405]: pgmap v491: 196 pgs: 32 creating+peering, 164 active+clean; 455 KiB data, 525 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:39.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:38 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:39.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:38 vm09 ceph-mon[54793]: osdmap e360: 8 total, 8 up, 8 in 2026-03-10T11:38:39.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:38 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:39.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:38 vm06 ceph-mon[49228]: osdmap e360: 8 total, 8 up, 8 in 2026-03-10T11:38:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:38 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:39.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:38 vm06 ceph-mon[57405]: osdmap e360: 8 total, 8 up, 8 in 2026-03-10T11:38:40.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:39 vm09 ceph-mon[54793]: osdmap e361: 8 total, 8 up, 8 in 2026-03-10T11:38:40.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:39 vm09 ceph-mon[54793]: pgmap v494: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 525 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:40.230 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:39 vm09 ceph-mon[54793]: osdmap e362: 8 total, 8 up, 8 in 2026-03-10T11:38:40.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:39 vm06 ceph-mon[49228]: osdmap e361: 8 total, 8 up, 8 in 2026-03-10T11:38:40.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:39 vm06 ceph-mon[49228]: pgmap v494: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 525 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:40.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:39 vm06 ceph-mon[49228]: osdmap e362: 8 total, 8 up, 8 in 2026-03-10T11:38:40.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:39 vm06 ceph-mon[57405]: osdmap e361: 8 total, 8 up, 8 in 2026-03-10T11:38:40.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:39 vm06 ceph-mon[57405]: pgmap v494: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 525 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:40.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:39 vm06 ceph-mon[57405]: osdmap e362: 8 total, 8 up, 8 in 2026-03-10T11:38:40.953 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestObject::test_seek PASSED [ 92%] 2026-03-10T11:38:40.994 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:38:40 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:38:40] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:38:41.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:40 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:41.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:40 vm06 ceph-mon[49228]: osdmap e363: 8 total, 8 up, 8 in 2026-03-10T11:38:41.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:40 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:41.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:40 vm06 ceph-mon[57405]: osdmap e363: 8 total, 8 up, 8 in 2026-03-10T11:38:41.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:41 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:41.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:41 vm09 ceph-mon[54793]: osdmap e363: 8 total, 8 up, 8 in 2026-03-10T11:38:42.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:42 vm06 ceph-mon[49228]: pgmap v497: 164 pgs: 164 active+clean; 455 KiB data, 526 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:42.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:42 vm06 ceph-mon[49228]: osdmap e364: 8 total, 8 up, 8 in 2026-03-10T11:38:42.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:42 vm06 ceph-mon[57405]: pgmap v497: 164 pgs: 164 active+clean; 455 KiB data, 526 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:42.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:42 vm06 ceph-mon[57405]: osdmap e364: 8 total, 8 up, 8 in 2026-03-10T11:38:42.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:42 vm09 ceph-mon[54793]: pgmap v497: 164 pgs: 164 active+clean; 455 KiB data, 526 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:42.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:42 vm09 ceph-mon[54793]: osdmap e364: 8 total, 8 up, 8 in 2026-03-10T11:38:43.975 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestObject::test_write PASSED [ 93%] 2026-03-10T11:38:44.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:43 vm06 ceph-mon[49228]: osdmap e365: 8 total, 8 up, 8 in 2026-03-10T11:38:44.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:43 vm06 ceph-mon[49228]: pgmap v500: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 526 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:44.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:43 vm06 ceph-mon[49228]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:44.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:43 vm06 ceph-mon[57405]: osdmap e365: 8 total, 8 up, 8 in 2026-03-10T11:38:44.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:43 vm06 ceph-mon[57405]: pgmap v500: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 526 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:44.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:43 vm06 ceph-mon[57405]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:44.281 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:43 vm09 ceph-mon[54793]: osdmap e365: 8 total, 8 up, 8 in 2026-03-10T11:38:44.281 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:43 vm09 ceph-mon[54793]: pgmap v500: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 526 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:44.281 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:43 vm09 ceph-mon[54793]: Health check update: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:45.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:45 vm06 ceph-mon[49228]: osdmap e366: 8 total, 8 up, 8 in 2026-03-10T11:38:45.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:45 vm06 ceph-mon[57405]: osdmap e366: 8 total, 8 up, 8 in 2026-03-10T11:38:45.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:45 vm09 ceph-mon[54793]: osdmap e366: 8 total, 8 up, 8 in 2026-03-10T11:38:46.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:46 vm06 ceph-mon[49228]: osdmap e367: 8 total, 8 up, 8 in 2026-03-10T11:38:46.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:46 vm06 ceph-mon[49228]: pgmap v503: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 526 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:46.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:46 vm06 ceph-mon[49228]: osdmap e368: 8 total, 8 up, 8 in 2026-03-10T11:38:46.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:46 vm06 ceph-mon[57405]: osdmap e367: 8 total, 8 up, 8 in 2026-03-10T11:38:46.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:46 vm06 ceph-mon[57405]: pgmap v503: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 526 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:46.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:46 vm06 ceph-mon[57405]: osdmap e368: 8 total, 8 up, 8 in 2026-03-10T11:38:46.361 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:46 vm09 ceph-mon[54793]: osdmap e367: 8 total, 8 up, 8 in 2026-03-10T11:38:46.361 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:46 vm09 ceph-mon[54793]: pgmap v503: 196 pgs: 32 unknown, 164 active+clean; 455 KiB data, 526 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:46.361 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:46 vm09 ceph-mon[54793]: osdmap e368: 8 total, 8 up, 8 in 2026-03-10T11:38:46.729 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:38:46 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:38:47.280 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:47 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:47.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:47 vm06 ceph-mon[49228]: osdmap e369: 8 total, 8 up, 8 in 2026-03-10T11:38:47.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:47 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:47.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:47 vm06 ceph-mon[57405]: osdmap e369: 8 total, 8 up, 8 in 2026-03-10T11:38:47.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:47 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:47.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:47 vm09 ceph-mon[54793]: osdmap e369: 8 total, 8 up, 8 in 2026-03-10T11:38:48.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:48 vm09 ceph-mon[54793]: pgmap v506: 196 pgs: 196 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:38:48.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:48 vm09 ceph-mon[54793]: osdmap e370: 8 total, 8 up, 8 in 2026-03-10T11:38:48.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:48 vm06 ceph-mon[49228]: pgmap v506: 196 pgs: 196 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:38:48.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:48 vm06 ceph-mon[49228]: osdmap e370: 8 total, 8 up, 8 in 2026-03-10T11:38:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:48 vm06 ceph-mon[57405]: pgmap v506: 196 pgs: 196 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:38:48.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:48 vm06 ceph-mon[57405]: osdmap e370: 8 total, 8 up, 8 in 2026-03-10T11:38:49.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:49 vm09 ceph-mon[54793]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:49.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:49 vm09 ceph-mon[54793]: osdmap e371: 8 total, 8 up, 8 in 2026-03-10T11:38:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:49 vm06 ceph-mon[49228]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:49.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:49 vm06 ceph-mon[49228]: osdmap e371: 8 total, 8 up, 8 in 2026-03-10T11:38:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:49 vm06 ceph-mon[57405]: Health check update: 2 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:49.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:49 vm06 ceph-mon[57405]: osdmap e371: 8 total, 8 up, 8 in 2026-03-10T11:38:50.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:50 vm09 ceph-mon[54793]: pgmap v509: 196 pgs: 196 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:38:50.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:50 vm09 ceph-mon[54793]: osdmap e372: 8 total, 8 up, 8 in 2026-03-10T11:38:50.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:50 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/993658212' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:50.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:50 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:50.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:50 vm06 ceph-mon[49228]: pgmap v509: 196 pgs: 196 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:38:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:50 vm06 ceph-mon[49228]: osdmap e372: 8 total, 8 up, 8 in 2026-03-10T11:38:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:50 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/993658212' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:50.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:50 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:50 vm06 ceph-mon[57405]: pgmap v509: 196 pgs: 196 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 255 B/s wr, 1 op/s 2026-03-10T11:38:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:50 vm06 ceph-mon[57405]: osdmap e372: 8 total, 8 up, 8 in 2026-03-10T11:38:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:50 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/993658212' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:50.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:50 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd unset", "key": "noup"}]: dispatch 2026-03-10T11:38:51.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:38:50 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:38:50] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:38:52.020 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestIoCtxSelfManagedSnaps::test PASSED [ 94%] 2026-03-10T11:38:52.037 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestCommand::test_monmap_dump PASSED [ 95%] 2026-03-10T11:38:52.049 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestCommand::test_osd_bench PASSED [ 96%] 2026-03-10T11:38:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:52 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:52 vm06 ceph-mon[49228]: osdmap e373: 8 total, 8 up, 8 in 2026-03-10T11:38:52.281 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:52 vm06 ceph-mon[49228]: pgmap v512: 196 pgs: 196 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.7 KiB/s rd, 511 B/s wr, 2 op/s 2026-03-10T11:38:52.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:52 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:52.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:52 vm06 ceph-mon[57405]: osdmap e373: 8 total, 8 up, 8 in 2026-03-10T11:38:52.281 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:52 vm06 ceph-mon[57405]: pgmap v512: 196 pgs: 196 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.7 KiB/s rd, 511 B/s wr, 2 op/s 2026-03-10T11:38:52.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:52 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd unset", "key": "noup"}]': finished 2026-03-10T11:38:52.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:52 vm09 ceph-mon[54793]: osdmap e373: 8 total, 8 up, 8 in 2026-03-10T11:38:52.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:52 vm09 ceph-mon[54793]: pgmap v512: 196 pgs: 196 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.7 KiB/s rd, 511 B/s wr, 2 op/s 2026-03-10T11:38:53.036 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestCommand::test_ceph_osd_pool_create_utf8 PASSED [ 97%] 2026-03-10T11:38:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:53 vm09 ceph-mon[54793]: osdmap e374: 8 total, 8 up, 8 in 2026-03-10T11:38:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:53 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1325238845' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-10T11:38:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:53 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1325238845' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:38:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:53 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1325238845' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json", "epoch": 1003}]: dispatch 2026-03-10T11:38:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:53 vm09 ceph-mon[54793]: from='client.? 192.168.123.106:0/1325411202' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 16, "pool": "\u9ec5"}]: dispatch 2026-03-10T11:38:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:53 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 16, "pool": "\u9ec5"}]: dispatch 2026-03-10T11:38:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:53 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:38:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:53 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:38:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:53 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:38:53.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:53 vm09 ceph-mon[54793]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[49228]: osdmap e374: 8 total, 8 up, 8 in 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1325238845' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1325238845' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1325238845' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json", "epoch": 1003}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[49228]: from='client.? 192.168.123.106:0/1325411202' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 16, "pool": "\u9ec5"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 16, "pool": "\u9ec5"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[49228]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[57405]: osdmap e374: 8 total, 8 up, 8 in 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1325238845' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1325238845' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1325238845' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json", "epoch": 1003}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[57405]: from='client.? 192.168.123.106:0/1325411202' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 16, "pool": "\u9ec5"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pg_num": 16, "pool": "\u9ec5"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T11:38:53.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:53 vm06 ceph-mon[57405]: from='mgr.24587 ' entity='mgr.y' 2026-03-10T11:38:54.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:54 vm09 ceph-mon[54793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pg_num": 16, "pool": "\u9ec5"}]': finished 2026-03-10T11:38:54.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:54 vm09 ceph-mon[54793]: osdmap e375: 8 total, 8 up, 8 in 2026-03-10T11:38:54.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:54 vm09 ceph-mon[54793]: pgmap v515: 180 pgs: 16 unknown, 164 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:54.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:54 vm09 ceph-mon[54793]: osdmap e376: 8 total, 8 up, 8 in 2026-03-10T11:38:54.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:54 vm06 ceph-mon[49228]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pg_num": 16, "pool": "\u9ec5"}]': finished 2026-03-10T11:38:54.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:54 vm06 ceph-mon[49228]: osdmap e375: 8 total, 8 up, 8 in 2026-03-10T11:38:54.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:54 vm06 ceph-mon[49228]: pgmap v515: 180 pgs: 16 unknown, 164 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:54.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:54 vm06 ceph-mon[49228]: osdmap e376: 8 total, 8 up, 8 in 2026-03-10T11:38:54.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:54 vm06 ceph-mon[57405]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pg_num": 16, "pool": "\u9ec5"}]': finished 2026-03-10T11:38:54.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:54 vm06 ceph-mon[57405]: osdmap e375: 8 total, 8 up, 8 in 2026-03-10T11:38:54.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:54 vm06 ceph-mon[57405]: pgmap v515: 180 pgs: 16 unknown, 164 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:38:54.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:54 vm06 ceph-mon[57405]: osdmap e376: 8 total, 8 up, 8 in 2026-03-10T11:38:56.364 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:56 vm09 ceph-mon[54793]: osdmap e377: 8 total, 8 up, 8 in 2026-03-10T11:38:56.364 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:56 vm09 ceph-mon[54793]: pgmap v518: 212 pgs: 48 unknown, 164 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:56.364 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:56 vm09 ceph-mon[54793]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:56.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:56 vm06 ceph-mon[49228]: osdmap e377: 8 total, 8 up, 8 in 2026-03-10T11:38:56.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:56 vm06 ceph-mon[49228]: pgmap v518: 212 pgs: 48 unknown, 164 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:56.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:56 vm06 ceph-mon[49228]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:56.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:56 vm06 ceph-mon[57405]: osdmap e377: 8 total, 8 up, 8 in 2026-03-10T11:38:56.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:56 vm06 ceph-mon[57405]: pgmap v518: 212 pgs: 48 unknown, 164 active+clean; 455 KiB data, 527 MiB used, 159 GiB / 160 GiB avail 2026-03-10T11:38:56.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:56 vm06 ceph-mon[57405]: from='mgr.24587 192.168.123.106:0/1678885686' entity='mgr.y' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T11:38:56.730 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:38:56 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug there is no tcmu-runner data available 2026-03-10T11:38:57.196 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestWatchNotify::test PASSED [ 98%] 2026-03-10T11:38:57.479 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:57 vm09 ceph-mon[54793]: Health check update: 3 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:57.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:57 vm09 ceph-mon[54793]: osdmap e378: 8 total, 8 up, 8 in 2026-03-10T11:38:57.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:57 vm09 ceph-mon[54793]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:57.530 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:57 vm06 ceph-mon[49228]: Health check update: 3 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:57.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:57 vm06 ceph-mon[49228]: osdmap e378: 8 total, 8 up, 8 in 2026-03-10T11:38:57.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:57 vm06 ceph-mon[49228]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:57.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:57 vm06 ceph-mon[57405]: Health check update: 3 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T11:38:57.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:57 vm06 ceph-mon[57405]: osdmap e378: 8 total, 8 up, 8 in 2026-03-10T11:38:57.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:57 vm06 ceph-mon[57405]: from='client.24440 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T11:38:58.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:58 vm09 ceph-mon[54793]: pgmap v520: 212 pgs: 212 active+clean; 455 KiB data, 528 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 254 B/s wr, 1 op/s 2026-03-10T11:38:58.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:58 vm09 ceph-mon[54793]: osdmap e379: 8 total, 8 up, 8 in 2026-03-10T11:38:58.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:58 vm06 ceph-mon[49228]: pgmap v520: 212 pgs: 212 active+clean; 455 KiB data, 528 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 254 B/s wr, 1 op/s 2026-03-10T11:38:58.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:58 vm06 ceph-mon[49228]: osdmap e379: 8 total, 8 up, 8 in 2026-03-10T11:38:58.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:58 vm06 ceph-mon[57405]: pgmap v520: 212 pgs: 212 active+clean; 455 KiB data, 528 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 254 B/s wr, 1 op/s 2026-03-10T11:38:58.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:58 vm06 ceph-mon[57405]: osdmap e379: 8 total, 8 up, 8 in 2026-03-10T11:38:59.480 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:38:59 vm09 ceph-mon[54793]: osdmap e380: 8 total, 8 up, 8 in 2026-03-10T11:38:59.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:38:59 vm06 ceph-mon[49228]: osdmap e380: 8 total, 8 up, 8 in 2026-03-10T11:38:59.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:38:59 vm06 ceph-mon[57405]: osdmap e380: 8 total, 8 up, 8 in 2026-03-10T11:39:00.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:39:00 vm06 ceph-mon[49228]: pgmap v523: 212 pgs: 32 unknown, 180 active+clean; 455 KiB data, 528 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:39:00.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:39:00 vm06 ceph-mon[49228]: osdmap e381: 8 total, 8 up, 8 in 2026-03-10T11:39:00.531 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:39:00 vm06 ceph-mon[49228]: osdmap e382: 8 total, 8 up, 8 in 2026-03-10T11:39:00.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:39:00 vm06 ceph-mon[57405]: pgmap v523: 212 pgs: 32 unknown, 180 active+clean; 455 KiB data, 528 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:39:00.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:39:00 vm06 ceph-mon[57405]: osdmap e381: 8 total, 8 up, 8 in 2026-03-10T11:39:00.531 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:39:00 vm06 ceph-mon[57405]: osdmap e382: 8 total, 8 up, 8 in 2026-03-10T11:39:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:39:00 vm09 ceph-mon[54793]: pgmap v523: 212 pgs: 32 unknown, 180 active+clean; 455 KiB data, 528 MiB used, 159 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T11:39:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:39:00 vm09 ceph-mon[54793]: osdmap e381: 8 total, 8 up, 8 in 2026-03-10T11:39:00.730 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:39:00 vm09 ceph-mon[54793]: osdmap e382: 8 total, 8 up, 8 in 2026-03-10T11:39:01.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:39:00 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: ::ffff:192.168.123.109 - - [10/Mar/2026:11:39:00] "GET /metrics HTTP/1.1" 503 1621 "" "Prometheus/2.51.0" 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py::TestWatchNotify::test_aio_notify PASSED [100%] 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout:=============================== warnings summary =============================== 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py:210 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: /home/ubuntu/cephtest/clone.client.0/src/test/pybind/test_rados.py:210: DeprecationWarning: invalid escape sequence \- 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: assert re.match('[0-9a-f\-]{36}', fsid, re.I) 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py:960 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: /home/ubuntu/cephtest/clone.client.0/src/test/pybind/test_rados.py:960: PytestUnknownMarkWarning: Unknown pytest.mark.wait - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/mark.html 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: @pytest.mark.wait 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py:996 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: /home/ubuntu/cephtest/clone.client.0/src/test/pybind/test_rados.py:996: PytestUnknownMarkWarning: Unknown pytest.mark.wait - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/mark.html 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: @pytest.mark.wait 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout:../../../clone.client.0/src/test/pybind/test_rados.py:1024 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: /home/ubuntu/cephtest/clone.client.0/src/test/pybind/test_rados.py:1024: PytestUnknownMarkWarning: Unknown pytest.mark.wait - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/mark.html 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: @pytest.mark.wait 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout::210 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout::210 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout::210 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout::210 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout::210 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout::210 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout::210 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout::210 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout::210 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: :210: DeprecationWarning: invalid escape sequence \- 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout: 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout:-- Docs: https://docs.pytest.org/en/stable/warnings.html 2026-03-10T11:39:01.260 INFO:tasks.workunit.client.0.vm06.stdout:================= 91 passed, 13 warnings in 330.91s (0:05:30) ================== 2026-03-10T11:39:01.272 INFO:tasks.workunit.client.0.vm06.stderr:+ exit 0 2026-03-10T11:39:01.273 INFO:teuthology.orchestra.run:Running command with timeout 3600 2026-03-10T11:39:01.273 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -rf -- /home/ubuntu/cephtest/mnt.0/client.0/tmp 2026-03-10T11:39:01.305 INFO:tasks.workunit:Stopping ['rados/test_python.sh'] on client.0... 2026-03-10T11:39:01.305 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -rf -- /home/ubuntu/cephtest/workunits.list.client.0 /home/ubuntu/cephtest/clone.client.0 2026-03-10T11:39:01.695 DEBUG:teuthology.parallel:result is None 2026-03-10T11:39:01.695 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -rf -- /home/ubuntu/cephtest/mnt.0/client.0 2026-03-10T11:39:01.718 INFO:tasks.workunit:Deleted dir /home/ubuntu/cephtest/mnt.0/client.0 2026-03-10T11:39:01.718 DEBUG:teuthology.orchestra.run.vm06:> rmdir -- /home/ubuntu/cephtest/mnt.0 2026-03-10T11:39:01.773 INFO:tasks.workunit:Deleted artificial mount point /home/ubuntu/cephtest/mnt.0/client.0 2026-03-10T11:39:01.773 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-10T11:39:01.775 INFO:tasks.cephadm:Teardown begin 2026-03-10T11:39:01.775 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T11:39:01.835 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T11:39:01.861 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-10T11:39:01.861 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 -- ceph mgr module disable cephadm 2026-03-10T11:39:02.039 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/mon.a/config 2026-03-10T11:39:02.058 INFO:teuthology.orchestra.run.vm06.stderr:Error: statfs /etc/ceph/ceph.client.admin.keyring: no such file or directory 2026-03-10T11:39:02.080 DEBUG:teuthology.orchestra.run:got remote process result: 125 2026-03-10T11:39:02.080 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-10T11:39:02.080 DEBUG:teuthology.orchestra.run.vm06:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T11:39:02.094 DEBUG:teuthology.orchestra.run.vm09:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T11:39:02.108 INFO:tasks.cephadm:Stopping all daemons... 2026-03-10T11:39:02.108 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-10T11:39:02.108 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.a 2026-03-10T11:39:02.385 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:39:02 vm06 systemd[1]: Stopping Ceph mon.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:02.385 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:39:02 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a[49224]: 2026-03-10T11:39:02.240+0000 7fbab4b38640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T11:39:02.385 INFO:journalctl@ceph.mon.a.vm06.stdout:Mar 10 11:39:02 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-a[49224]: 2026-03-10T11:39:02.240+0000 7fbab4b38640 -1 mon.a@0(leader) e3 *** Got Signal Terminated *** 2026-03-10T11:39:02.483 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.a.service' 2026-03-10T11:39:02.514 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:02.514 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-10T11:39:02.514 INFO:tasks.cephadm.mon.b:Stopping mon.c... 2026-03-10T11:39:02.514 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.c 2026-03-10T11:39:02.642 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 10 11:39:02 vm06 systemd[1]: Stopping Ceph mon.c for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:02.822 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.c.service' 2026-03-10T11:39:02.850 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:02.850 INFO:tasks.cephadm.mon.b:Stopped mon.c 2026-03-10T11:39:02.850 INFO:tasks.cephadm.mon.b:Stopping mon.b... 2026-03-10T11:39:02.850 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.b 2026-03-10T11:39:03.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:39:02 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:39:02] ENGINE Bus STOPPING 2026-03-10T11:39:03.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:39:02 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:39:02] ENGINE HTTP Server cherrypy._cpwsgi_server.CPWSGIServer(('::', 9283)) shut down 2026-03-10T11:39:03.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:39:02 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:39:02] ENGINE Bus STOPPED 2026-03-10T11:39:03.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:39:02 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:39:02] ENGINE Bus STARTING 2026-03-10T11:39:03.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:39:02 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:39:02] ENGINE Serving on http://:::9283 2026-03-10T11:39:03.031 INFO:journalctl@ceph.mgr.y.vm06.stdout:Mar 10 11:39:02 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mgr-y[49435]: [10/Mar/2026:11:39:02] ENGINE Bus STARTED 2026-03-10T11:39:03.121 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:39:02 vm09 systemd[1]: Stopping Ceph mon.b for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:03.121 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:39:02 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-b[54789]: 2026-03-10T11:39:02.955+0000 7f0810681640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T11:39:03.121 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:39:02 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-b[54789]: 2026-03-10T11:39:02.955+0000 7f0810681640 -1 mon.b@1(peon) e3 *** Got Signal Terminated *** 2026-03-10T11:39:03.121 INFO:journalctl@ceph.mon.b.vm09.stdout:Mar 10 11:39:03 vm09 podman[83384]: 2026-03-10 11:39:03.111849651 +0000 UTC m=+0.174569130 container died 63b259ab8cee880ef64e79e1d1f1d1a0f2f25003ad421c083e3d6f35dea98fc9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-mon-b, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, io.buildah.version=1.41.3, CEPH_REF=squid) 2026-03-10T11:39:03.183 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mon.b.service' 2026-03-10T11:39:03.216 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:03.216 INFO:tasks.cephadm.mon.b:Stopped mon.b 2026-03-10T11:39:03.216 INFO:tasks.cephadm.mgr.y:Stopping mgr.y... 2026-03-10T11:39:03.217 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mgr.y 2026-03-10T11:39:03.444 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mgr.y.service' 2026-03-10T11:39:03.472 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:03.472 INFO:tasks.cephadm.mgr.y:Stopped mgr.y 2026-03-10T11:39:03.472 INFO:tasks.cephadm.mgr.x:Stopping mgr.x... 2026-03-10T11:39:03.472 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mgr.x 2026-03-10T11:39:03.689 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@mgr.x.service' 2026-03-10T11:39:03.718 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:03.718 INFO:tasks.cephadm.mgr.x:Stopped mgr.x 2026-03-10T11:39:03.718 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-10T11:39:03.718 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.0 2026-03-10T11:39:04.031 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:03 vm06 systemd[1]: Stopping Ceph osd.0 for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:04.031 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:03 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0[61331]: 2026-03-10T11:39:03.808+0000 7f8b99a9a640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T11:39:04.031 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:03 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0[61331]: 2026-03-10T11:39:03.808+0000 7f8b99a9a640 -1 osd.0 383 *** Got signal Terminated *** 2026-03-10T11:39:04.031 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:03 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0[61331]: 2026-03-10T11:39:03.808+0000 7f8b99a9a640 -1 osd.0 383 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T11:39:09.148 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:08 vm06 podman[91325]: 2026-03-10 11:39:08.847573109 +0000 UTC m=+5.049578408 container died 6443f2c4eea0bb1872fb04a30501fb1a883a75391c581afeda742c9ff19c153f (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-10T11:39:09.149 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:08 vm06 podman[91325]: 2026-03-10 11:39:08.859719365 +0000 UTC m=+5.061724655 container remove 6443f2c4eea0bb1872fb04a30501fb1a883a75391c581afeda742c9ff19c153f (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, ceph=True, org.label-schema.license=GPLv2) 2026-03-10T11:39:09.149 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:08 vm06 bash[91325]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0 2026-03-10T11:39:09.149 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:08 vm06 podman[91393]: 2026-03-10 11:39:08.990712288 +0000 UTC m=+0.017376577 container create cc7b79073251f6166cfbdf3947992c72ea403902ebcfcd57c870423425e273c2 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0-deactivate, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True) 2026-03-10T11:39:09.149 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:09 vm06 podman[91393]: 2026-03-10 11:39:09.026316275 +0000 UTC m=+0.052980564 container init cc7b79073251f6166cfbdf3947992c72ea403902ebcfcd57c870423425e273c2 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0-deactivate, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, OSD_FLAVOR=default, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, CEPH_REF=squid) 2026-03-10T11:39:09.149 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:09 vm06 podman[91393]: 2026-03-10 11:39:09.029032559 +0000 UTC m=+0.055696848 container start cc7b79073251f6166cfbdf3947992c72ea403902ebcfcd57c870423425e273c2 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0-deactivate, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid) 2026-03-10T11:39:09.149 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:09 vm06 podman[91393]: 2026-03-10 11:39:09.029959445 +0000 UTC m=+0.056623734 container attach cc7b79073251f6166cfbdf3947992c72ea403902ebcfcd57c870423425e273c2 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0-deactivate, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True) 2026-03-10T11:39:09.149 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:09 vm06 podman[91393]: 2026-03-10 11:39:08.982502752 +0000 UTC m=+0.009167041 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T11:39:09.149 INFO:journalctl@ceph.osd.0.vm06.stdout:Mar 10 11:39:09 vm06 podman[91393]: 2026-03-10 11:39:09.148727234 +0000 UTC m=+0.175391513 container died cc7b79073251f6166cfbdf3947992c72ea403902ebcfcd57c870423425e273c2 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-0-deactivate, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, CEPH_REF=squid) 2026-03-10T11:39:09.173 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.0.service' 2026-03-10T11:39:09.204 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:09.204 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-10T11:39:09.204 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-10T11:39:09.204 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.1 2026-03-10T11:39:09.532 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:09 vm06 systemd[1]: Stopping Ceph osd.1 for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:09.532 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:09 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1[66565]: 2026-03-10T11:39:09.345+0000 7f2f5387c640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T11:39:09.532 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:09 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1[66565]: 2026-03-10T11:39:09.345+0000 7f2f5387c640 -1 osd.1 383 *** Got signal Terminated *** 2026-03-10T11:39:09.532 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:09 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1[66565]: 2026-03-10T11:39:09.345+0000 7f2f5387c640 -1 osd.1 383 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T11:39:14.693 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:14 vm06 podman[91491]: 2026-03-10 11:39:14.371475596 +0000 UTC m=+5.036827702 container died b7323517468e4d99ec419cb5430cecb7948e0e8fbe727c983169f1ecbff37e1e (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, OSD_FLAVOR=default, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid) 2026-03-10T11:39:14.693 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:14 vm06 podman[91491]: 2026-03-10 11:39:14.394728282 +0000 UTC m=+5.060080398 container remove b7323517468e4d99ec419cb5430cecb7948e0e8fbe727c983169f1ecbff37e1e (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T11:39:14.693 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:14 vm06 bash[91491]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1 2026-03-10T11:39:14.693 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:14 vm06 podman[91572]: 2026-03-10 11:39:14.518509195 +0000 UTC m=+0.013800844 container create 8a564d3fa09efd9738f69839dbc086e3c13d21aa54d45725009dedd2348f979d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1-deactivate, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, CEPH_REF=squid, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True) 2026-03-10T11:39:14.693 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:14 vm06 podman[91572]: 2026-03-10 11:39:14.559280824 +0000 UTC m=+0.054572484 container init 8a564d3fa09efd9738f69839dbc086e3c13d21aa54d45725009dedd2348f979d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1-deactivate, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, OSD_FLAVOR=default) 2026-03-10T11:39:14.693 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:14 vm06 podman[91572]: 2026-03-10 11:39:14.562304755 +0000 UTC m=+0.057596415 container start 8a564d3fa09efd9738f69839dbc086e3c13d21aa54d45725009dedd2348f979d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1-deactivate, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, io.buildah.version=1.41.3) 2026-03-10T11:39:14.693 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:14 vm06 podman[91572]: 2026-03-10 11:39:14.564330116 +0000 UTC m=+0.059621776 container attach 8a564d3fa09efd9738f69839dbc086e3c13d21aa54d45725009dedd2348f979d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, CEPH_REF=squid) 2026-03-10T11:39:14.693 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:14 vm06 podman[91572]: 2026-03-10 11:39:14.512765764 +0000 UTC m=+0.008057434 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T11:39:14.693 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 10 11:39:14 vm06 podman[91572]: 2026-03-10 11:39:14.692010578 +0000 UTC m=+0.187302249 container died 8a564d3fa09efd9738f69839dbc086e3c13d21aa54d45725009dedd2348f979d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-1-deactivate, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) 2026-03-10T11:39:14.716 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.1.service' 2026-03-10T11:39:14.746 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:14.746 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-10T11:39:14.746 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-10T11:39:14.746 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.2 2026-03-10T11:39:15.031 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:39:14 vm06 systemd[1]: Stopping Ceph osd.2 for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:15.031 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:39:14 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2[71586]: 2026-03-10T11:39:14.883+0000 7f4b57f41640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T11:39:15.031 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:39:14 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2[71586]: 2026-03-10T11:39:14.883+0000 7f4b57f41640 -1 osd.2 383 *** Got signal Terminated *** 2026-03-10T11:39:15.031 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:39:14 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2[71586]: 2026-03-10T11:39:14.883+0000 7f4b57f41640 -1 osd.2 383 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T11:39:20.169 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:39:19 vm06 podman[91667]: 2026-03-10 11:39:19.915158657 +0000 UTC m=+5.046334547 container died 7c3a4bb265e35b7bb2c473376be8c8a0a7cf7cc527e8c33c647736d48c018189 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, CEPH_REF=squid, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0) 2026-03-10T11:39:20.169 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:39:19 vm06 podman[91667]: 2026-03-10 11:39:19.933363745 +0000 UTC m=+5.064539635 container remove 7c3a4bb265e35b7bb2c473376be8c8a0a7cf7cc527e8c33c647736d48c018189 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, io.buildah.version=1.41.3, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid) 2026-03-10T11:39:20.169 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:39:19 vm06 bash[91667]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2 2026-03-10T11:39:20.169 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:39:20 vm06 podman[91736]: 2026-03-10 11:39:20.077408711 +0000 UTC m=+0.018348208 container create 97a4d3e155af73fb225c07f569bb9235385b013b6cf2dd05aec6412aad006d93 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2-deactivate, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T11:39:20.169 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:39:20 vm06 podman[91736]: 2026-03-10 11:39:20.116057786 +0000 UTC m=+0.056997283 container init 97a4d3e155af73fb225c07f569bb9235385b013b6cf2dd05aec6412aad006d93 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2-deactivate, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, org.label-schema.schema-version=1.0) 2026-03-10T11:39:20.169 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:39:20 vm06 podman[91736]: 2026-03-10 11:39:20.120938102 +0000 UTC m=+0.061877599 container start 97a4d3e155af73fb225c07f569bb9235385b013b6cf2dd05aec6412aad006d93 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2-deactivate, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-10T11:39:20.169 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 11:39:20 vm06 podman[91736]: 2026-03-10 11:39:20.122273521 +0000 UTC m=+0.063213009 container attach 97a4d3e155af73fb225c07f569bb9235385b013b6cf2dd05aec6412aad006d93 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-2-deactivate, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-10T11:39:20.294 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.2.service' 2026-03-10T11:39:20.331 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:20.331 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-10T11:39:20.331 INFO:tasks.cephadm.osd.3:Stopping osd.3... 2026-03-10T11:39:20.331 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.3 2026-03-10T11:39:20.470 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:39:20 vm06 systemd[1]: Stopping Ceph osd.3 for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:20.781 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:39:20 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3[76780]: 2026-03-10T11:39:20.469+0000 7ff886e64640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.3 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T11:39:20.781 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:39:20 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3[76780]: 2026-03-10T11:39:20.470+0000 7ff886e64640 -1 osd.3 383 *** Got signal Terminated *** 2026-03-10T11:39:20.781 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:39:20 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3[76780]: 2026-03-10T11:39:20.470+0000 7ff886e64640 -1 osd.3 383 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T11:39:25.770 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:39:25 vm06 podman[91831]: 2026-03-10 11:39:25.507394095 +0000 UTC m=+5.050233686 container died 8ac3345d610e3493c6e36ce47b787fa9244298f116c1012c3a4b217695af52d9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3, org.label-schema.vendor=CentOS, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, CEPH_REF=squid, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3) 2026-03-10T11:39:25.770 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:39:25 vm06 podman[91831]: 2026-03-10 11:39:25.533737327 +0000 UTC m=+5.076576918 container remove 8ac3345d610e3493c6e36ce47b787fa9244298f116c1012c3a4b217695af52d9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, CEPH_REF=squid, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, ceph=True, org.label-schema.build-date=20260223) 2026-03-10T11:39:25.770 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:39:25 vm06 bash[91831]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3 2026-03-10T11:39:25.770 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:39:25 vm06 podman[91901]: 2026-03-10 11:39:25.677410666 +0000 UTC m=+0.015634677 container create 5e177ab4e7670a3f8fe94fc894f03a945de849996515180abea15e5f8af4326d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3-deactivate, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, io.buildah.version=1.41.3, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, CEPH_REF=squid) 2026-03-10T11:39:25.770 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:39:25 vm06 podman[91901]: 2026-03-10 11:39:25.715036057 +0000 UTC m=+0.053260077 container init 5e177ab4e7670a3f8fe94fc894f03a945de849996515180abea15e5f8af4326d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3-deactivate, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, CEPH_REF=squid, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2) 2026-03-10T11:39:25.770 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:39:25 vm06 podman[91901]: 2026-03-10 11:39:25.718002619 +0000 UTC m=+0.056226630 container start 5e177ab4e7670a3f8fe94fc894f03a945de849996515180abea15e5f8af4326d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3-deactivate, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default) 2026-03-10T11:39:25.770 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 11:39:25 vm06 podman[91901]: 2026-03-10 11:39:25.71897501 +0000 UTC m=+0.057199021 container attach 5e177ab4e7670a3f8fe94fc894f03a945de849996515180abea15e5f8af4326d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-3-deactivate, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, CEPH_REF=squid) 2026-03-10T11:39:25.873 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.3.service' 2026-03-10T11:39:25.903 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:25.903 INFO:tasks.cephadm.osd.3:Stopped osd.3 2026-03-10T11:39:25.903 INFO:tasks.cephadm.osd.4:Stopping osd.4... 2026-03-10T11:39:25.903 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.4 2026-03-10T11:39:26.230 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:39:25 vm09 systemd[1]: Stopping Ceph osd.4 for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:26.230 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:39:26 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4[59293]: 2026-03-10T11:39:26.005+0000 7f530fcfa640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.4 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T11:39:26.230 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:39:26 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4[59293]: 2026-03-10T11:39:26.005+0000 7f530fcfa640 -1 osd.4 383 *** Got signal Terminated *** 2026-03-10T11:39:26.230 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:39:26 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4[59293]: 2026-03-10T11:39:26.005+0000 7f530fcfa640 -1 osd.4 383 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T11:39:29.230 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:28 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:39:28.929+0000 7f57876b7640 -1 osd.5 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.576141+0000 front 2026-03-10T11:39:06.576125+0000 (oldest deadline 2026-03-10T11:39:28.875892+0000) 2026-03-10T11:39:30.229 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:29 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:39:29.944+0000 7f57876b7640 -1 osd.5 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.576141+0000 front 2026-03-10T11:39:06.576125+0000 (oldest deadline 2026-03-10T11:39:28.875892+0000) 2026-03-10T11:39:30.954 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:30 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:30.629+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:31.213 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:30 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:39:30.954+0000 7f57876b7640 -1 osd.5 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.576141+0000 front 2026-03-10T11:39:06.576125+0000 (oldest deadline 2026-03-10T11:39:28.875892+0000) 2026-03-10T11:39:31.213 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:39:31 vm09 podman[83597]: 2026-03-10 11:39:31.033211322 +0000 UTC m=+5.041822436 container died 4330ed0f09a2a43875f30843d42db6dc876f7503eeaef59ddde637b1b3de2fdc (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, ceph=True, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T11:39:31.213 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:39:31 vm09 podman[83597]: 2026-03-10 11:39:31.066926352 +0000 UTC m=+5.075537466 container remove 4330ed0f09a2a43875f30843d42db6dc876f7503eeaef59ddde637b1b3de2fdc (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3) 2026-03-10T11:39:31.213 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 11:39:31 vm09 bash[83597]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-4 2026-03-10T11:39:31.404 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.4.service' 2026-03-10T11:39:31.433 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:31.433 INFO:tasks.cephadm.osd.4:Stopped osd.4 2026-03-10T11:39:31.433 INFO:tasks.cephadm.osd.5:Stopping osd.5... 2026-03-10T11:39:31.434 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.5 2026-03-10T11:39:31.979 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:31 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:31.676+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:31.980 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:31 vm09 systemd[1]: Stopping Ceph osd.5 for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:31.980 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:31 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:39:31.570+0000 7f578b09e640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.5 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T11:39:31.980 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:31 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:39:31.570+0000 7f578b09e640 -1 osd.5 383 *** Got signal Terminated *** 2026-03-10T11:39:31.980 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:31 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:39:31.570+0000 7f578b09e640 -1 osd.5 383 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T11:39:32.479 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:31 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:39:31.988+0000 7f57876b7640 -1 osd.5 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.576141+0000 front 2026-03-10T11:39:06.576125+0000 (oldest deadline 2026-03-10T11:39:28.875892+0000) 2026-03-10T11:39:32.979 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:32 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:32.711+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:32.980 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:32 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:32.961+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:33.479 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:33 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:39:33.019+0000 7f57876b7640 -1 osd.5 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.576141+0000 front 2026-03-10T11:39:06.576125+0000 (oldest deadline 2026-03-10T11:39:28.875892+0000) 2026-03-10T11:39:33.925 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:33 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:33.725+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:34.229 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:34 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:39:34.026+0000 7f57876b7640 -1 osd.5 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.576141+0000 front 2026-03-10T11:39:06.576125+0000 (oldest deadline 2026-03-10T11:39:28.875892+0000) 2026-03-10T11:39:34.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:33 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:33.925+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:35.068 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:34 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:34.747+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:35.068 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:34 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:34.949+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:35.479 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:39:35.068+0000 7f57876b7640 -1 osd.5 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.576141+0000 front 2026-03-10T11:39:06.576125+0000 (oldest deadline 2026-03-10T11:39:28.875892+0000) 2026-03-10T11:39:36.076 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:35.745+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:36.076 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:35 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:35.971+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:36.479 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5[64240]: 2026-03-10T11:39:36.076+0000 7f57876b7640 -1 osd.5 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.576141+0000 front 2026-03-10T11:39:06.576125+0000 (oldest deadline 2026-03-10T11:39:28.875892+0000) 2026-03-10T11:39:36.895 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:36 vm09 podman[83770]: 2026-03-10 11:39:36.591392936 +0000 UTC m=+5.033418890 container died 29782b359c89f332ebd52d061cf654c70bafbe4c50c4ccc1b9ea8252d25b04c2 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default) 2026-03-10T11:39:36.895 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:36 vm09 podman[83770]: 2026-03-10 11:39:36.620921629 +0000 UTC m=+5.062947583 container remove 29782b359c89f332ebd52d061cf654c70bafbe4c50c4ccc1b9ea8252d25b04c2 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True) 2026-03-10T11:39:36.895 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:36 vm09 bash[83770]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5 2026-03-10T11:39:36.895 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:36 vm09 podman[83840]: 2026-03-10 11:39:36.741241781 +0000 UTC m=+0.013932481 container create aef20c59fb1f7848a8267148e33e9e8a2f1ecc28cc39f0ed025816e6d16ff322 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5-deactivate, org.label-schema.license=GPLv2, OSD_FLAVOR=default, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, io.buildah.version=1.41.3) 2026-03-10T11:39:36.895 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:36 vm09 podman[83840]: 2026-03-10 11:39:36.77739154 +0000 UTC m=+0.050082258 container init aef20c59fb1f7848a8267148e33e9e8a2f1ecc28cc39f0ed025816e6d16ff322 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5-deactivate, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.label-schema.vendor=CentOS) 2026-03-10T11:39:36.895 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:36 vm09 podman[83840]: 2026-03-10 11:39:36.780147979 +0000 UTC m=+0.052838688 container start aef20c59fb1f7848a8267148e33e9e8a2f1ecc28cc39f0ed025816e6d16ff322 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5-deactivate, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T11:39:36.895 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:36 vm09 podman[83840]: 2026-03-10 11:39:36.789617334 +0000 UTC m=+0.062308054 container attach aef20c59fb1f7848a8267148e33e9e8a2f1ecc28cc39f0ed025816e6d16ff322 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-5-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, CEPH_REF=squid, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS) 2026-03-10T11:39:36.895 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 11:39:36 vm09 podman[83840]: 2026-03-10 11:39:36.735286914 +0000 UTC m=+0.007977634 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T11:39:36.896 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:36.716+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:36.923 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.5.service' 2026-03-10T11:39:36.953 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:36.953 INFO:tasks.cephadm.osd.5:Stopped osd.5 2026-03-10T11:39:36.953 INFO:tasks.cephadm.osd.6:Stopping osd.6... 2026-03-10T11:39:36.953 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.6 2026-03-10T11:39:37.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:37 vm09 systemd[1]: Stopping Ceph osd.6 for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:37.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:37.088+0000 7fcb195ec640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.6 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T11:39:37.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:37.088+0000 7fcb195ec640 -1 osd.6 383 *** Got signal Terminated *** 2026-03-10T11:39:37.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:37.088+0000 7fcb195ec640 -1 osd.6 383 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T11:39:37.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:36 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:36.935+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:38.229 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:37.764+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:38.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:37.968+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:38.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:37 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:37.968+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.541047+0000 front 2026-03-10T11:39:12.541005+0000 (oldest deadline 2026-03-10T11:39:37.840688+0000) 2026-03-10T11:39:39.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:38 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:38.978+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:39.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:38 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:38.978+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.541047+0000 front 2026-03-10T11:39:12.541005+0000 (oldest deadline 2026-03-10T11:39:37.840688+0000) 2026-03-10T11:39:39.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:38 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:38.745+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:39.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:38 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:38.745+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.707049+0000 front 2026-03-10T11:39:12.707119+0000 (oldest deadline 2026-03-10T11:39:38.606675+0000) 2026-03-10T11:39:40.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:39.783+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:40.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:39.783+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.707049+0000 front 2026-03-10T11:39:12.707119+0000 (oldest deadline 2026-03-10T11:39:38.606675+0000) 2026-03-10T11:39:40.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:39.944+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:40.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:39 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:39.944+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.541047+0000 front 2026-03-10T11:39:12.541005+0000 (oldest deadline 2026-03-10T11:39:37.840688+0000) 2026-03-10T11:39:41.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:40.804+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:41.230 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:40.804+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.707049+0000 front 2026-03-10T11:39:12.707119+0000 (oldest deadline 2026-03-10T11:39:38.606675+0000) 2026-03-10T11:39:41.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:40.983+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:41.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:40.983+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.541047+0000 front 2026-03-10T11:39:12.541005+0000 (oldest deadline 2026-03-10T11:39:37.840688+0000) 2026-03-10T11:39:41.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:40 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:40.983+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6822 osd.2 since back 2026-03-10T11:39:18.941465+0000 front 2026-03-10T11:39:18.941245+0000 (oldest deadline 2026-03-10T11:39:40.641095+0000) 2026-03-10T11:39:42.126 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:41 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:41.854+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:06.906532+0000 front 2026-03-10T11:39:06.906614+0000 (oldest deadline 2026-03-10T11:39:30.406205+0000) 2026-03-10T11:39:42.126 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:41 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:41.854+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.707049+0000 front 2026-03-10T11:39:12.707119+0000 (oldest deadline 2026-03-10T11:39:38.606675+0000) 2026-03-10T11:39:42.126 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:41 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6[69445]: 2026-03-10T11:39:41.854+0000 7fcb15404640 -1 osd.6 383 heartbeat_check: no reply from 192.168.123.106:6822 osd.2 since back 2026-03-10T11:39:18.607175+0000 front 2026-03-10T11:39:18.607081+0000 (oldest deadline 2026-03-10T11:39:40.906860+0000) 2026-03-10T11:39:42.126 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:41 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:41.981+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:42.126 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:41 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:41.981+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.541047+0000 front 2026-03-10T11:39:12.541005+0000 (oldest deadline 2026-03-10T11:39:37.840688+0000) 2026-03-10T11:39:42.126 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:41 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:41.981+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6822 osd.2 since back 2026-03-10T11:39:18.941465+0000 front 2026-03-10T11:39:18.941245+0000 (oldest deadline 2026-03-10T11:39:40.641095+0000) 2026-03-10T11:39:42.430 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:42 vm09 podman[83935]: 2026-03-10 11:39:42.126672976 +0000 UTC m=+5.050819196 container died e709f2dab0e3ca0dc5d0de77d664ef9c779dbccb0b047d912de497a416a57b65 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, ceph=True, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T11:39:42.430 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:42 vm09 podman[83935]: 2026-03-10 11:39:42.152469334 +0000 UTC m=+5.076615554 container remove e709f2dab0e3ca0dc5d0de77d664ef9c779dbccb0b047d912de497a416a57b65 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6, ceph=True, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T11:39:42.430 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:42 vm09 bash[83935]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6 2026-03-10T11:39:42.430 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:42 vm09 podman[84000]: 2026-03-10 11:39:42.275892335 +0000 UTC m=+0.014983758 container create dd4183f8c4d580ebe487ffc4f3a4d192bdc8f840aa8729e49c9b95daca08ae30 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6-deactivate, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, CEPH_REF=squid, ceph=True, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3) 2026-03-10T11:39:42.430 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:42 vm09 podman[84000]: 2026-03-10 11:39:42.304560007 +0000 UTC m=+0.043651430 container init dd4183f8c4d580ebe487ffc4f3a4d192bdc8f840aa8729e49c9b95daca08ae30 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6-deactivate, OSD_FLAVOR=default, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True) 2026-03-10T11:39:42.430 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:42 vm09 podman[84000]: 2026-03-10 11:39:42.307212462 +0000 UTC m=+0.046303885 container start dd4183f8c4d580ebe487ffc4f3a4d192bdc8f840aa8729e49c9b95daca08ae30 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6-deactivate, ceph=True, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-10T11:39:42.430 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:42 vm09 podman[84000]: 2026-03-10 11:39:42.308225398 +0000 UTC m=+0.047316821 container attach dd4183f8c4d580ebe487ffc4f3a4d192bdc8f840aa8729e49c9b95daca08ae30 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-6-deactivate, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T11:39:42.430 INFO:journalctl@ceph.osd.6.vm09.stdout:Mar 10 11:39:42 vm09 podman[84000]: 2026-03-10 11:39:42.269786806 +0000 UTC m=+0.008878240 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T11:39:42.458 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.6.service' 2026-03-10T11:39:42.488 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:42.488 INFO:tasks.cephadm.osd.6:Stopped osd.6 2026-03-10T11:39:42.488 INFO:tasks.cephadm.osd.7:Stopping osd.7... 2026-03-10T11:39:42.488 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.7 2026-03-10T11:39:42.730 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:42 vm09 systemd[1]: Stopping Ceph osd.7 for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:42.730 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:42 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:42.623+0000 7f5098623640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.7 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T11:39:42.730 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:42 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:42.623+0000 7f5098623640 -1 osd.7 383 *** Got signal Terminated *** 2026-03-10T11:39:42.730 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:42 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:42.623+0000 7f5098623640 -1 osd.7 383 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T11:39:43.229 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:42 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:42.949+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:43.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:42 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:42.949+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.541047+0000 front 2026-03-10T11:39:12.541005+0000 (oldest deadline 2026-03-10T11:39:37.840688+0000) 2026-03-10T11:39:43.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:42 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:42.949+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6822 osd.2 since back 2026-03-10T11:39:18.941465+0000 front 2026-03-10T11:39:18.941245+0000 (oldest deadline 2026-03-10T11:39:40.641095+0000) 2026-03-10T11:39:44.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:43 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:43.926+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:44.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:43 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:43.926+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.541047+0000 front 2026-03-10T11:39:12.541005+0000 (oldest deadline 2026-03-10T11:39:37.840688+0000) 2026-03-10T11:39:44.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:43 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:43.926+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6822 osd.2 since back 2026-03-10T11:39:18.941465+0000 front 2026-03-10T11:39:18.941245+0000 (oldest deadline 2026-03-10T11:39:40.641095+0000) 2026-03-10T11:39:45.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:44 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:44.962+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:45.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:44 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:44.962+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.541047+0000 front 2026-03-10T11:39:12.541005+0000 (oldest deadline 2026-03-10T11:39:37.840688+0000) 2026-03-10T11:39:45.230 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:44 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:44.962+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6822 osd.2 since back 2026-03-10T11:39:18.941465+0000 front 2026-03-10T11:39:18.941245+0000 (oldest deadline 2026-03-10T11:39:40.641095+0000) 2026-03-10T11:39:46.480 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:46 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:46.006+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:46.480 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:46 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:46.006+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.541047+0000 front 2026-03-10T11:39:12.541005+0000 (oldest deadline 2026-03-10T11:39:37.840688+0000) 2026-03-10T11:39:46.480 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:46 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:46.006+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6822 osd.2 since back 2026-03-10T11:39:18.941465+0000 front 2026-03-10T11:39:18.941245+0000 (oldest deadline 2026-03-10T11:39:40.641095+0000) 2026-03-10T11:39:47.480 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:47 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:47.038+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6806 osd.0 since back 2026-03-10T11:39:07.840995+0000 front 2026-03-10T11:39:07.840771+0000 (oldest deadline 2026-03-10T11:39:32.540444+0000) 2026-03-10T11:39:47.480 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:47 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:47.038+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6814 osd.1 since back 2026-03-10T11:39:12.541047+0000 front 2026-03-10T11:39:12.541005+0000 (oldest deadline 2026-03-10T11:39:37.840688+0000) 2026-03-10T11:39:47.480 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:47 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7[74677]: 2026-03-10T11:39:47.038+0000 7f509443b640 -1 osd.7 383 heartbeat_check: no reply from 192.168.123.106:6822 osd.2 since back 2026-03-10T11:39:18.941465+0000 front 2026-03-10T11:39:18.941245+0000 (oldest deadline 2026-03-10T11:39:40.641095+0000) 2026-03-10T11:39:47.976 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:47 vm09 podman[84095]: 2026-03-10 11:39:47.658833948 +0000 UTC m=+5.047062464 container died e61c09e4966f3000a6c435d47ea96822d41e3760903cc8f60e6f85c9d13e21c8 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True) 2026-03-10T11:39:47.976 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:47 vm09 podman[84095]: 2026-03-10 11:39:47.678324142 +0000 UTC m=+5.066552648 container remove e61c09e4966f3000a6c435d47ea96822d41e3760903cc8f60e6f85c9d13e21c8 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid) 2026-03-10T11:39:47.976 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:47 vm09 bash[84095]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7 2026-03-10T11:39:47.976 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:47 vm09 podman[84162]: 2026-03-10 11:39:47.801219462 +0000 UTC m=+0.014696291 container create 5d9b9359cbf1ddcde254ff1503de1148acbc5af4d4b1ff4c546b2fe5ee43d3ef (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7-deactivate, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-10T11:39:47.976 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:47 vm09 podman[84162]: 2026-03-10 11:39:47.840581496 +0000 UTC m=+0.054058325 container init 5d9b9359cbf1ddcde254ff1503de1148acbc5af4d4b1ff4c546b2fe5ee43d3ef (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7-deactivate, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.label-schema.license=GPLv2, ceph=True, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default) 2026-03-10T11:39:47.976 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:47 vm09 podman[84162]: 2026-03-10 11:39:47.843060857 +0000 UTC m=+0.056537686 container start 5d9b9359cbf1ddcde254ff1503de1148acbc5af4d4b1ff4c546b2fe5ee43d3ef (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7-deactivate, OSD_FLAVOR=default, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223) 2026-03-10T11:39:47.976 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:47 vm09 podman[84162]: 2026-03-10 11:39:47.843978996 +0000 UTC m=+0.057455825 container attach 5d9b9359cbf1ddcde254ff1503de1148acbc5af4d4b1ff4c546b2fe5ee43d3ef (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-osd-7-deactivate, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.3, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T11:39:47.976 INFO:journalctl@ceph.osd.7.vm09.stdout:Mar 10 11:39:47 vm09 podman[84162]: 2026-03-10 11:39:47.795401683 +0000 UTC m=+0.008878512 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T11:39:48.002 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@osd.7.service' 2026-03-10T11:39:48.031 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:48.031 INFO:tasks.cephadm.osd.7:Stopped osd.7 2026-03-10T11:39:48.031 INFO:tasks.cephadm.ceph.rgw.foo.a:Stopping rgw.foo.a... 2026-03-10T11:39:48.031 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@rgw.foo.a 2026-03-10T11:39:48.489 INFO:journalctl@ceph.rgw.foo.a.vm06.stdout:Mar 10 11:39:48 vm06 systemd[1]: Stopping Ceph rgw.foo.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:39:48.489 INFO:journalctl@ceph.rgw.foo.a.vm06.stdout:Mar 10 11:39:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-rgw-foo-a[81067]: 2026-03-10T11:39:48.137+0000 7f3afd326640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/radosgw -n client.rgw.foo.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T11:39:48.489 INFO:journalctl@ceph.rgw.foo.a.vm06.stdout:Mar 10 11:39:48 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-rgw-foo-a[81067]: 2026-03-10T11:39:48.137+0000 7f3b00b95980 -1 shutting down 2026-03-10T11:39:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:39:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:39:48.656Z caller=refresh.go:90 level=error component="discovery manager scrape" discovery=http config=node msg="Unable to refresh target groups" err="Get \"http://192.168.123.106:8765/sd/prometheus/sd-config?service=node-exporter\": dial tcp 192.168.123.106:8765: connect: connection refused" 2026-03-10T11:39:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:39:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:39:48.656Z caller=refresh.go:90 level=error component="discovery manager scrape" discovery=http config=nfs msg="Unable to refresh target groups" err="Get \"http://192.168.123.106:8765/sd/prometheus/sd-config?service=nfs\": dial tcp 192.168.123.106:8765: connect: connection refused" 2026-03-10T11:39:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:39:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:39:48.660Z caller=refresh.go:90 level=error component="discovery manager scrape" discovery=http config=ceph msg="Unable to refresh target groups" err="Get \"http://192.168.123.106:8765/sd/prometheus/sd-config?service=mgr-prometheus\": dial tcp 192.168.123.106:8765: connect: connection refused" 2026-03-10T11:39:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:39:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:39:48.660Z caller=refresh.go:90 level=error component="discovery manager notify" discovery=http config=config-0 msg="Unable to refresh target groups" err="Get \"http://192.168.123.106:8765/sd/prometheus/sd-config?service=alertmanager\": dial tcp 192.168.123.106:8765: connect: connection refused" 2026-03-10T11:39:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:39:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:39:48.660Z caller=refresh.go:90 level=error component="discovery manager scrape" discovery=http config=nvmeof msg="Unable to refresh target groups" err="Get \"http://192.168.123.106:8765/sd/prometheus/sd-config?service=nvmeof\": dial tcp 192.168.123.106:8765: connect: connection refused" 2026-03-10T11:39:48.980 INFO:journalctl@ceph.prometheus.a.vm09.stdout:Mar 10 11:39:48 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-prometheus-a[81913]: ts=2026-03-10T11:39:48.660Z caller=refresh.go:90 level=error component="discovery manager scrape" discovery=http config=ceph-exporter msg="Unable to refresh target groups" err="Get \"http://192.168.123.106:8765/sd/prometheus/sd-config?service=ceph-exporter\": dial tcp 192.168.123.106:8765: connect: connection refused" 2026-03-10T11:39:58.237 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@rgw.foo.a.service' 2026-03-10T11:39:58.265 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:58.265 INFO:tasks.cephadm.ceph.rgw.foo.a:Stopped rgw.foo.a 2026-03-10T11:39:58.265 INFO:tasks.cephadm.prometheus.a:Stopping prometheus.a... 2026-03-10T11:39:58.265 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@prometheus.a 2026-03-10T11:39:58.445 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@prometheus.a.service' 2026-03-10T11:39:58.473 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T11:39:58.473 INFO:tasks.cephadm.prometheus.a:Stopped prometheus.a 2026-03-10T11:39:58.473 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm rm-cluster --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 --force --keep-logs 2026-03-10T11:39:58.591 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:40:00.101 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:39:59 vm06 systemd[1]: Stopping Ceph alertmanager.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:40:00.101 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:40:00 vm06 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a[87776]: ts=2026-03-10T11:40:00.056Z caller=main.go:583 level=info msg="Received SIGTERM, exiting gracefully..." 2026-03-10T11:40:00.101 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:40:00 vm06 podman[92520]: 2026-03-10 11:40:00.067053275 +0000 UTC m=+0.023263026 container died d067b405a2f9af608fc12e8da0d517c85068cd4673120d32c246b1e119cd414d (image=quay.io/prometheus/alertmanager:v0.25.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a, maintainer=The Prometheus Authors ) 2026-03-10T11:40:00.101 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:40:00 vm06 podman[92520]: 2026-03-10 11:40:00.080599482 +0000 UTC m=+0.036809223 container remove d067b405a2f9af608fc12e8da0d517c85068cd4673120d32c246b1e119cd414d (image=quay.io/prometheus/alertmanager:v0.25.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a, maintainer=The Prometheus Authors ) 2026-03-10T11:40:00.101 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:40:00 vm06 podman[92520]: 2026-03-10 11:40:00.081628268 +0000 UTC m=+0.037838019 volume remove 9941b3ef7aacf5bbfcf28fe6b4c2b11321afebb37d561514847bbf20a14e4bbb 2026-03-10T11:40:00.101 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:40:00 vm06 bash[92520]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-alertmanager-a 2026-03-10T11:40:00.355 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:40:00 vm06 systemd[1]: Stopping Ceph node-exporter.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:40:00.355 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:40:00 vm06 podman[92623]: 2026-03-10 11:40:00.351936095 +0000 UTC m=+0.013644722 container stop 2e000946ad6d135b7409d27a9ce0a2256b01388f61b9ef719a26be6c508cdcf9 (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a, maintainer=The Prometheus Authors ) 2026-03-10T11:40:00.355 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:40:00 vm06 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@alertmanager.a.service: Deactivated successfully. 2026-03-10T11:40:00.355 INFO:journalctl@ceph.alertmanager.a.vm06.stdout:Mar 10 11:40:00 vm06 systemd[1]: Stopped Ceph alertmanager.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:40:00.658 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:40:00 vm06 podman[92623]: 2026-03-10 11:40:00.356409489 +0000 UTC m=+0.018118126 container died 2e000946ad6d135b7409d27a9ce0a2256b01388f61b9ef719a26be6c508cdcf9 (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a, maintainer=The Prometheus Authors ) 2026-03-10T11:40:00.658 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:40:00 vm06 podman[92623]: 2026-03-10 11:40:00.368081348 +0000 UTC m=+0.029789964 container remove 2e000946ad6d135b7409d27a9ce0a2256b01388f61b9ef719a26be6c508cdcf9 (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a, maintainer=The Prometheus Authors ) 2026-03-10T11:40:00.658 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:40:00 vm06 bash[92623]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-a 2026-03-10T11:40:00.659 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:40:00 vm06 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@node-exporter.a.service: Main process exited, code=exited, status=143/n/a 2026-03-10T11:40:00.659 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:40:00 vm06 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@node-exporter.a.service: Failed with result 'exit-code'. 2026-03-10T11:40:00.659 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:40:00 vm06 systemd[1]: Stopped Ceph node-exporter.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:40:00.659 INFO:journalctl@ceph.node-exporter.a.vm06.stdout:Mar 10 11:40:00 vm06 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@node-exporter.a.service: Consumed 1.051s CPU time. 2026-03-10T11:40:00.958 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm rm-cluster --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 --force --keep-logs 2026-03-10T11:40:01.095 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:40:02.429 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:40:02 vm09 systemd[1]: Stopping Ceph iscsi.iscsi.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:40:02.429 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:40:02 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a[78954]: debug Shutdown received 2026-03-10T11:40:12.566 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:40:12 vm09 bash[84688]: time="2026-03-10T11:40:12Z" level=warning msg="StopSignal SIGTERM failed to stop container ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a in 10 seconds, resorting to SIGKILL" 2026-03-10T11:40:12.566 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:40:12 vm09 podman[84688]: 2026-03-10 11:40:12.324928109 +0000 UTC m=+10.036207850 container died 42354789c474e6bce60aeb219bc9d5059ca7937e6e9960ef4a0a84851a99b781 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2) 2026-03-10T11:40:12.566 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:40:12 vm09 podman[84688]: 2026-03-10 11:40:12.344096931 +0000 UTC m=+10.055376662 container remove 42354789c474e6bce60aeb219bc9d5059ca7937e6e9960ef4a0a84851a99b781 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, ceph=True, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, OSD_FLAVOR=default) 2026-03-10T11:40:12.566 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:40:12 vm09 bash[84688]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-iscsi-iscsi-a 2026-03-10T11:40:12.566 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:40:12 vm09 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@iscsi.iscsi.a.service: Main process exited, code=exited, status=137/n/a 2026-03-10T11:40:12.566 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:40:12 vm09 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@iscsi.iscsi.a.service: Failed with result 'exit-code'. 2026-03-10T11:40:12.566 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:40:12 vm09 systemd[1]: Stopped Ceph iscsi.iscsi.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:40:12.566 INFO:journalctl@ceph.iscsi.iscsi.a.vm09.stdout:Mar 10 11:40:12 vm09 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@iscsi.iscsi.a.service: Consumed 1.173s CPU time. 2026-03-10T11:40:13.094 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:12 vm09 systemd[1]: Stopping Ceph grafana.a for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:40:13.094 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:13 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=server t=2026-03-10T11:40:13.054466177Z level=info msg="Shutdown started" reason="System signal: terminated" 2026-03-10T11:40:13.094 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:13 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=ticker t=2026-03-10T11:40:13.054908035Z level=info msg=stopped last_tick=2026-03-10T11:40:10Z 2026-03-10T11:40:13.094 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:13 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=grafana-apiserver t=2026-03-10T11:40:13.054932791Z level=info msg="StorageObjectCountTracker pruner is exiting" 2026-03-10T11:40:13.094 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:13 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=tracing t=2026-03-10T11:40:13.054944232Z level=info msg="Closing tracing" 2026-03-10T11:40:13.094 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:13 vm09 ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a[81017]: logger=sqlstore.transactions t=2026-03-10T11:40:13.065515892Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=0 code="database is locked" 2026-03-10T11:40:13.094 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:13 vm09 podman[84932]: 2026-03-10 11:40:13.074802476 +0000 UTC m=+0.032983764 container died affeafdf638bc1051dd0ce4943c9d40f1f98b582f0657130cc9d85e63b42c3ac (image=quay.io/ceph/grafana:10.4.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a, maintainer=Grafana Labs ) 2026-03-10T11:40:13.355 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:40:13 vm09 systemd[1]: Stopping Ceph node-exporter.b for 52119272-1c74-11f1-990e-e7cc1c0e3258... 2026-03-10T11:40:13.355 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:13 vm09 podman[84932]: 2026-03-10 11:40:13.09489653 +0000 UTC m=+0.053077818 container remove affeafdf638bc1051dd0ce4943c9d40f1f98b582f0657130cc9d85e63b42c3ac (image=quay.io/ceph/grafana:10.4.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a, maintainer=Grafana Labs ) 2026-03-10T11:40:13.356 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:13 vm09 bash[84932]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-grafana-a 2026-03-10T11:40:13.356 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:13 vm09 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@grafana.a.service: Deactivated successfully. 2026-03-10T11:40:13.356 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:13 vm09 systemd[1]: Stopped Ceph grafana.a for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:40:13.356 INFO:journalctl@ceph.grafana.a.vm09.stdout:Mar 10 11:40:13 vm09 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@grafana.a.service: Consumed 3.269s CPU time. 2026-03-10T11:40:13.670 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:40:13 vm09 podman[85030]: 2026-03-10 11:40:13.35626921 +0000 UTC m=+0.011682390 container stop 3b0b72eff720abfb5b984cc56d49c55c2819e6b5d04e50b28ac6360ca1725dfd (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b, maintainer=The Prometheus Authors ) 2026-03-10T11:40:13.671 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:40:13 vm09 podman[85030]: 2026-03-10 11:40:13.385662691 +0000 UTC m=+0.041075871 container died 3b0b72eff720abfb5b984cc56d49c55c2819e6b5d04e50b28ac6360ca1725dfd (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b, maintainer=The Prometheus Authors ) 2026-03-10T11:40:13.671 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:40:13 vm09 podman[85030]: 2026-03-10 11:40:13.396872246 +0000 UTC m=+0.052285426 container remove 3b0b72eff720abfb5b984cc56d49c55c2819e6b5d04e50b28ac6360ca1725dfd (image=quay.io/prometheus/node-exporter:v1.7.0, name=ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b, maintainer=The Prometheus Authors ) 2026-03-10T11:40:13.671 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:40:13 vm09 bash[85030]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258-node-exporter-b 2026-03-10T11:40:13.671 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:40:13 vm09 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@node-exporter.b.service: Main process exited, code=exited, status=143/n/a 2026-03-10T11:40:13.671 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:40:13 vm09 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@node-exporter.b.service: Failed with result 'exit-code'. 2026-03-10T11:40:13.671 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:40:13 vm09 systemd[1]: Stopped Ceph node-exporter.b for 52119272-1c74-11f1-990e-e7cc1c0e3258. 2026-03-10T11:40:13.671 INFO:journalctl@ceph.node-exporter.b.vm09.stdout:Mar 10 11:40:13 vm09 systemd[1]: ceph-52119272-1c74-11f1-990e-e7cc1c0e3258@node-exporter.b.service: Consumed 1.015s CPU time. 2026-03-10T11:40:13.975 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T11:40:14.005 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T11:40:14.028 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-10T11:40:14.028 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1016/remote/vm06/crash 2026-03-10T11:40:14.028 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/crash -- . 2026-03-10T11:40:14.072 INFO:teuthology.orchestra.run.vm06.stderr:tar: /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/crash: Cannot open: No such file or directory 2026-03-10T11:40:14.072 INFO:teuthology.orchestra.run.vm06.stderr:tar: Error is not recoverable: exiting now 2026-03-10T11:40:14.073 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1016/remote/vm09/crash 2026-03-10T11:40:14.073 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/crash -- . 2026-03-10T11:40:14.097 INFO:teuthology.orchestra.run.vm09.stderr:tar: /var/lib/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/crash: Cannot open: No such file or directory 2026-03-10T11:40:14.097 INFO:teuthology.orchestra.run.vm09.stderr:tar: Error is not recoverable: exiting now 2026-03-10T11:40:14.098 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-10T11:40:14.098 DEBUG:teuthology.orchestra.run.vm06:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v 'but it is still running' | egrep -v 'overall HEALTH_' | egrep -v '\(OSDMAP_FLAGS\)' | egrep -v '\(PG_' | egrep -v '\(OSD_' | egrep -v '\(OBJECT_' | egrep -v '\(POOL_APP_NOT_ENABLED\)' | head -n 1 2026-03-10T11:40:14.143 INFO:tasks.cephadm:Compressing logs... 2026-03-10T11:40:14.143 DEBUG:teuthology.orchestra.run.vm06:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T11:40:14.185 DEBUG:teuthology.orchestra.run.vm09:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T11:40:14.206 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T11:40:14.206 INFO:teuthology.orchestra.run.vm06.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T11:40:14.207 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mon.a.log 2026-03-10T11:40:14.207 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.log 2026-03-10T11:40:14.208 INFO:teuthology.orchestra.run.vm09.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T11:40:14.208 INFO:teuthology.orchestra.run.vm09.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T11:40:14.208 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mon.a.log: 91.7% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T11:40:14.208 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mgr.y.log 2026-03-10T11:40:14.209 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-volume.log 2026-03-10T11:40:14.210 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.log: 92.4% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.log.gz 2026-03-10T11:40:14.210 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.audit.log 2026-03-10T11:40:14.211 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mon.b.log 2026-03-10T11:40:14.215 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-volume.log: 91.3% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T11:40:14.215 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.cephadm.log 2026-03-10T11:40:14.217 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mon.b.log: 95.5% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-volume.log.gz 2026-03-10T11:40:14.218 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.audit.log 2026-03-10T11:40:14.218 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.cephadm.log: 80.7% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.cephadm.log.gz 2026-03-10T11:40:14.219 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.log 2026-03-10T11:40:14.219 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mgr.y.log: gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.cephadm.log 2026-03-10T11:40:14.220 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.audit.log: 90.6% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.audit.log.gz 2026-03-10T11:40:14.220 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mgr.x.log 2026-03-10T11:40:14.221 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.audit.log: 94.2% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.audit.log.gz 2026-03-10T11:40:14.222 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.log: 86.5% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.log.gz 2026-03-10T11:40:14.222 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.4.log 2026-03-10T11:40:14.223 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-volume.log 2026-03-10T11:40:14.223 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.cephadm.log: 88.9% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph.cephadm.log.gz 2026-03-10T11:40:14.226 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mgr.x.log: 91.0% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mgr.x.log.gz 2026-03-10T11:40:14.226 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.5.log 2026-03-10T11:40:14.227 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mon.c.log 2026-03-10T11:40:14.231 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.4.log: gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.6.log 2026-03-10T11:40:14.237 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.0.log 2026-03-10T11:40:14.243 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mon.c.log: gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.1.log 2026-03-10T11:40:14.243 INFO:teuthology.orchestra.run.vm06.stderr: 95.4% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-volume.log.gz 2026-03-10T11:40:14.243 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.5.log: gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.7.log 2026-03-10T11:40:14.250 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.6.log: gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/tcmu-runner.log 2026-03-10T11:40:14.253 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.0.log: gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.2.log 2026-03-10T11:40:14.261 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.7.log: /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/tcmu-runner.log: 63.1% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/tcmu-runner.log.gz 2026-03-10T11:40:14.262 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.1.log: gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.3.log 2026-03-10T11:40:14.274 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.2.log: gzip -5 --verbose -- /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-client.rgw.foo.a.log 2026-03-10T11:40:14.283 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.3.log: /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-client.rgw.foo.a.log: 58.5% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-client.rgw.foo.a.log.gz 2026-03-10T11:40:14.469 INFO:teuthology.orchestra.run.vm06.stderr: 89.8% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mgr.y.log.gz 2026-03-10T11:40:14.573 INFO:teuthology.orchestra.run.vm09.stderr: 91.8% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mon.b.log.gz 2026-03-10T11:40:14.861 INFO:teuthology.orchestra.run.vm06.stderr: 92.0% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mon.c.log.gz 2026-03-10T11:40:15.205 INFO:teuthology.orchestra.run.vm06.stderr: 91.6% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-mon.a.log.gz 2026-03-10T11:40:16.566 INFO:teuthology.orchestra.run.vm09.stderr: 94.7% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.6.log.gz 2026-03-10T11:40:16.601 INFO:teuthology.orchestra.run.vm09.stderr: 94.7% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.5.log.gz 2026-03-10T11:40:16.647 INFO:teuthology.orchestra.run.vm06.stderr: 94.7% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.2.log.gz 2026-03-10T11:40:16.681 INFO:teuthology.orchestra.run.vm09.stderr: 94.7% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.7.log.gz 2026-03-10T11:40:16.689 INFO:teuthology.orchestra.run.vm09.stderr: 94.8% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.4.log.gz 2026-03-10T11:40:16.690 INFO:teuthology.orchestra.run.vm09.stderr: 2026-03-10T11:40:16.690 INFO:teuthology.orchestra.run.vm09.stderr:real 0m2.492s 2026-03-10T11:40:16.690 INFO:teuthology.orchestra.run.vm09.stderr:user 0m4.764s 2026-03-10T11:40:16.690 INFO:teuthology.orchestra.run.vm09.stderr:sys 0m0.197s 2026-03-10T11:40:16.842 INFO:teuthology.orchestra.run.vm06.stderr: 94.7% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.1.log.gz 2026-03-10T11:40:16.892 INFO:teuthology.orchestra.run.vm06.stderr: 94.8% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.0.log.gz 2026-03-10T11:40:17.058 INFO:teuthology.orchestra.run.vm06.stderr: 94.9% -- replaced with /var/log/ceph/52119272-1c74-11f1-990e-e7cc1c0e3258/ceph-osd.3.log.gz 2026-03-10T11:40:17.059 INFO:teuthology.orchestra.run.vm06.stderr: 2026-03-10T11:40:17.059 INFO:teuthology.orchestra.run.vm06.stderr:real 0m2.863s 2026-03-10T11:40:17.059 INFO:teuthology.orchestra.run.vm06.stderr:user 0m5.308s 2026-03-10T11:40:17.059 INFO:teuthology.orchestra.run.vm06.stderr:sys 0m0.239s 2026-03-10T11:40:17.060 INFO:tasks.cephadm:Archiving logs... 2026-03-10T11:40:17.060 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1016/remote/vm06/log 2026-03-10T11:40:17.060 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T11:40:17.378 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1016/remote/vm09/log 2026-03-10T11:40:17.378 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T11:40:17.629 INFO:tasks.cephadm:Removing cluster... 2026-03-10T11:40:17.629 DEBUG:teuthology.orchestra.run.vm06:> sudo cephadm rm-cluster --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 --force 2026-03-10T11:40:17.748 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:40:17.949 DEBUG:teuthology.orchestra.run.vm09:> sudo cephadm rm-cluster --fsid 52119272-1c74-11f1-990e-e7cc1c0e3258 --force 2026-03-10T11:40:18.071 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 52119272-1c74-11f1-990e-e7cc1c0e3258 2026-03-10T11:40:18.293 INFO:tasks.cephadm:Teardown complete 2026-03-10T11:40:18.293 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-10T11:40:18.295 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-10T11:40:18.295 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T11:40:18.297 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T11:40:18.329 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-10T11:40:18.329 DEBUG:teuthology.orchestra.run.vm06:> 2026-03-10T11:40:18.329 DEBUG:teuthology.orchestra.run.vm06:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-10T11:40:18.329 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y remove $d || true 2026-03-10T11:40:18.329 DEBUG:teuthology.orchestra.run.vm06:> done 2026-03-10T11:40:18.334 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-10T11:40:18.334 DEBUG:teuthology.orchestra.run.vm09:> 2026-03-10T11:40:18.334 DEBUG:teuthology.orchestra.run.vm09:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-10T11:40:18.334 DEBUG:teuthology.orchestra.run.vm09:> sudo yum -y remove $d || true 2026-03-10T11:40:18.334 DEBUG:teuthology.orchestra.run.vm09:> done 2026-03-10T11:40:18.530 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:18.530 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:18.530 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T11:40:18.530 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:18.530 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T11:40:18.530 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-10T11:40:18.530 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T11:40:18.530 INFO:teuthology.orchestra.run.vm09.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-10T11:40:18.531 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:18.531 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T11:40:18.531 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:18.531 INFO:teuthology.orchestra.run.vm09.stdout:Remove 2 Packages 2026-03-10T11:40:18.531 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:18.531 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 39 M 2026-03-10T11:40:18.531 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T11:40:18.531 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 39 M 2026-03-10T11:40:18.532 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T11:40:18.533 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T11:40:18.533 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T11:40:18.534 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T11:40:18.534 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T11:40:18.546 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T11:40:18.546 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T11:40:18.548 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T11:40:18.548 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T11:40:18.575 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T11:40:18.577 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T11:40:18.596 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:18.596 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:18.596 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T11:40:18.596 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-10T11:40:18.596 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-10T11:40:18.596 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:18.598 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:18.598 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:18.598 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T11:40:18.598 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-10T11:40:18.598 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-10T11:40:18.598 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:18.599 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:18.602 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:18.608 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:18.610 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:18.622 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T11:40:18.624 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T11:40:18.690 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T11:40:18.690 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:18.697 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T11:40:18.697 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:18.737 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T11:40:18.737 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:18.737 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T11:40:18.737 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-10T11:40:18.737 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:18.737 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:18.752 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T11:40:18.752 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:18.752 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T11:40:18.752 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-10T11:40:18.752 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:18.752 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:18.932 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout:Remove 4 Packages 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 212 M 2026-03-10T11:40:18.933 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T11:40:18.936 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T11:40:18.936 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T11:40:18.944 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:18.944 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:18.944 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T11:40:18.944 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:18.944 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T11:40:18.944 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-10T11:40:18.944 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T11:40:18.944 INFO:teuthology.orchestra.run.vm09.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-10T11:40:18.945 INFO:teuthology.orchestra.run.vm09.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-10T11:40:18.945 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-10T11:40:18.945 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:18.945 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T11:40:18.945 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:18.945 INFO:teuthology.orchestra.run.vm09.stdout:Remove 4 Packages 2026-03-10T11:40:18.945 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:18.945 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 212 M 2026-03-10T11:40:18.945 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T11:40:18.947 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T11:40:18.947 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T11:40:18.958 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T11:40:18.958 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T11:40:18.970 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T11:40:18.971 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T11:40:19.017 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T11:40:19.024 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T11:40:19.026 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-10T11:40:19.029 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-10T11:40:19.034 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T11:40:19.040 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T11:40:19.043 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-10T11:40:19.044 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T11:40:19.047 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-10T11:40:19.062 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T11:40:19.128 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T11:40:19.128 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T11:40:19.128 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-10T11:40:19.128 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-10T11:40:19.144 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T11:40:19.144 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T11:40:19.144 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-10T11:40:19.144 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-10T11:40:19.178 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-10T11:40:19.178 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:19.178 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T11:40:19.178 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-10T11:40:19.178 INFO:teuthology.orchestra.run.vm09.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T11:40:19.178 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:19.178 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:19.200 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-10T11:40:19.200 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:19.200 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T11:40:19.200 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-10T11:40:19.200 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T11:40:19.200 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:19.200 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:19.379 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout:Remove 8 Packages 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 28 M 2026-03-10T11:40:19.380 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T11:40:19.383 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T11:40:19.383 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T11:40:19.407 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T11:40:19.407 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T11:40:19.419 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T11:40:19.420 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:19.421 INFO:teuthology.orchestra.run.vm06.stdout:Remove 8 Packages 2026-03-10T11:40:19.421 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:19.421 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 28 M 2026-03-10T11:40:19.421 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T11:40:19.423 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T11:40:19.423 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T11:40:19.447 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T11:40:19.447 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T11:40:19.448 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T11:40:19.453 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T11:40:19.458 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-10T11:40:19.459 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-10T11:40:19.462 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-10T11:40:19.465 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-10T11:40:19.467 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-10T11:40:19.486 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T11:40:19.486 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:19.486 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T11:40:19.486 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-10T11:40:19.486 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-10T11:40:19.486 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:19.487 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T11:40:19.494 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T11:40:19.498 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T11:40:19.503 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T11:40:19.506 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-10T11:40:19.508 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-10T11:40:19.511 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-10T11:40:19.512 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T11:40:19.513 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:19.513 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T11:40:19.513 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-10T11:40:19.513 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-10T11:40:19.513 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:19.514 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-10T11:40:19.514 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T11:40:19.516 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-10T11:40:19.535 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T11:40:19.535 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:19.535 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T11:40:19.535 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-10T11:40:19.535 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-10T11:40:19.535 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:19.535 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T11:40:19.541 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T11:40:19.561 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T11:40:19.561 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:19.561 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T11:40:19.561 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-10T11:40:19.561 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-10T11:40:19.561 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:19.563 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T11:40:19.604 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T11:40:19.604 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T11:40:19.604 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-10T11:40:19.604 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-10T11:40:19.604 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-10T11:40:19.604 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-10T11:40:19.604 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-10T11:40:19.604 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-10T11:40:19.652 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T11:40:19.652 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T11:40:19.652 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-10T11:40:19.652 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-10T11:40:19.652 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-10T11:40:19.652 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-10T11:40:19.652 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-10T11:40:19.652 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout: zip-3.0-35.el9.x86_64 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:19.659 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:19.714 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:19.859 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:19.864 INFO:teuthology.orchestra.run.vm09.stdout:=========================================================================================== 2026-03-10T11:40:19.864 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T11:40:19.864 INFO:teuthology.orchestra.run.vm09.stdout:=========================================================================================== 2026-03-10T11:40:19.864 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout:Removing dependent packages: 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-10T11:40:19.865 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout:=========================================================================================== 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout:Remove 100 Packages 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 612 M 2026-03-10T11:40:19.866 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T11:40:19.892 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T11:40:19.892 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T11:40:19.990 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-10T11:40:19.995 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-10T11:40:19.996 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout:Remove 100 Packages 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 612 M 2026-03-10T11:40:19.997 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T11:40:20.022 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T11:40:20.022 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T11:40:20.122 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T11:40:20.122 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T11:40:20.134 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T11:40:20.134 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/100 2026-03-10T11:40:20.141 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/100 2026-03-10T11:40:20.160 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/100 2026-03-10T11:40:20.161 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:20.161 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T11:40:20.161 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-10T11:40:20.161 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-10T11:40:20.161 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:20.161 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/100 2026-03-10T11:40:20.173 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/100 2026-03-10T11:40:20.195 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/100 2026-03-10T11:40:20.196 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/100 2026-03-10T11:40:20.250 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/100 2026-03-10T11:40:20.258 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/100 2026-03-10T11:40:20.263 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/100 2026-03-10T11:40:20.263 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T11:40:20.263 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/100 2026-03-10T11:40:20.263 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/100 2026-03-10T11:40:20.271 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/100 2026-03-10T11:40:20.276 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/100 2026-03-10T11:40:20.283 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/100 2026-03-10T11:40:20.287 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/100 2026-03-10T11:40:20.289 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/100 2026-03-10T11:40:20.289 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:20.289 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T11:40:20.289 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-10T11:40:20.289 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-10T11:40:20.289 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:20.289 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/100 2026-03-10T11:40:20.296 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/100 2026-03-10T11:40:20.301 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/100 2026-03-10T11:40:20.302 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/100 2026-03-10T11:40:20.323 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/100 2026-03-10T11:40:20.323 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:20.323 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T11:40:20.323 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-10T11:40:20.323 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-10T11:40:20.323 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:20.324 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/100 2026-03-10T11:40:20.324 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/100 2026-03-10T11:40:20.328 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/100 2026-03-10T11:40:20.337 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/100 2026-03-10T11:40:20.352 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/100 2026-03-10T11:40:20.352 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:20.352 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T11:40:20.352 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:20.360 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/100 2026-03-10T11:40:20.377 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/100 2026-03-10T11:40:20.434 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/100 2026-03-10T11:40:20.434 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/100 2026-03-10T11:40:20.437 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/100 2026-03-10T11:40:20.438 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/100 2026-03-10T11:40:20.439 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/100 2026-03-10T11:40:20.442 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/100 2026-03-10T11:40:20.448 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/100 2026-03-10T11:40:20.450 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/100 2026-03-10T11:40:20.456 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/100 2026-03-10T11:40:20.456 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/100 2026-03-10T11:40:20.461 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/100 2026-03-10T11:40:20.469 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/100 2026-03-10T11:40:20.469 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/100 2026-03-10T11:40:20.474 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/100 2026-03-10T11:40:20.476 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/100 2026-03-10T11:40:20.486 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/100 2026-03-10T11:40:20.493 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/100 2026-03-10T11:40:20.494 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/100 2026-03-10T11:40:20.494 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:20.494 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T11:40:20.494 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-10T11:40:20.494 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-10T11:40:20.494 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:20.498 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/100 2026-03-10T11:40:20.507 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/100 2026-03-10T11:40:20.521 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/100 2026-03-10T11:40:20.522 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:20.522 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T11:40:20.522 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:20.522 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/100 2026-03-10T11:40:20.529 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/100 2026-03-10T11:40:20.530 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/100 2026-03-10T11:40:20.533 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/100 2026-03-10T11:40:20.537 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/100 2026-03-10T11:40:20.540 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/100 2026-03-10T11:40:20.542 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/100 2026-03-10T11:40:20.545 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/100 2026-03-10T11:40:20.549 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/100 2026-03-10T11:40:20.553 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/100 2026-03-10T11:40:20.554 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/100 2026-03-10T11:40:20.557 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/100 2026-03-10T11:40:20.562 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/100 2026-03-10T11:40:20.569 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/100 2026-03-10T11:40:20.574 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/100 2026-03-10T11:40:20.584 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/100 2026-03-10T11:40:20.589 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/100 2026-03-10T11:40:20.616 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/100 2026-03-10T11:40:20.623 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/100 2026-03-10T11:40:20.626 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/100 2026-03-10T11:40:20.634 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/100 2026-03-10T11:40:20.644 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/100 2026-03-10T11:40:20.644 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/100 2026-03-10T11:40:20.651 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/100 2026-03-10T11:40:20.655 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/100 2026-03-10T11:40:20.672 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/100 2026-03-10T11:40:20.684 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/100 2026-03-10T11:40:20.685 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-10T11:40:20.685 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:20.685 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/100 2026-03-10T11:40:20.712 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/100 2026-03-10T11:40:20.727 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/100 2026-03-10T11:40:20.733 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/100 2026-03-10T11:40:20.736 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/100 2026-03-10T11:40:20.739 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/100 2026-03-10T11:40:20.741 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/100 2026-03-10T11:40:20.756 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/100 2026-03-10T11:40:20.760 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/100 2026-03-10T11:40:20.760 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:20.760 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T11:40:20.760 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-10T11:40:20.760 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-10T11:40:20.760 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:20.761 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/100 2026-03-10T11:40:20.769 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/100 2026-03-10T11:40:20.769 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-10T11:40:20.769 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:20.770 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/100 2026-03-10T11:40:20.774 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/100 2026-03-10T11:40:20.779 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/100 2026-03-10T11:40:20.782 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/100 2026-03-10T11:40:20.785 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/100 2026-03-10T11:40:20.788 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/100 2026-03-10T11:40:20.792 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/100 2026-03-10T11:40:20.795 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/100 2026-03-10T11:40:20.797 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/100 2026-03-10T11:40:20.802 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/100 2026-03-10T11:40:20.810 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/100 2026-03-10T11:40:20.816 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/100 2026-03-10T11:40:20.818 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/100 2026-03-10T11:40:20.821 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/100 2026-03-10T11:40:20.839 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/100 2026-03-10T11:40:20.840 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:20.840 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T11:40:20.840 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-10T11:40:20.840 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-10T11:40:20.840 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:20.841 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/100 2026-03-10T11:40:20.848 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/100 2026-03-10T11:40:20.852 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/100 2026-03-10T11:40:20.857 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/100 2026-03-10T11:40:20.860 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/100 2026-03-10T11:40:20.861 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/100 2026-03-10T11:40:20.863 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/100 2026-03-10T11:40:20.865 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/100 2026-03-10T11:40:20.865 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/100 2026-03-10T11:40:20.868 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/100 2026-03-10T11:40:20.870 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/100 2026-03-10T11:40:20.872 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/100 2026-03-10T11:40:20.873 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/100 2026-03-10T11:40:20.877 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/100 2026-03-10T11:40:20.877 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/100 2026-03-10T11:40:20.880 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/100 2026-03-10T11:40:20.902 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/100 2026-03-10T11:40:20.902 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:20.902 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T11:40:20.902 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:20.903 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/100 2026-03-10T11:40:20.910 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/100 2026-03-10T11:40:20.912 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/100 2026-03-10T11:40:20.915 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/100 2026-03-10T11:40:20.918 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/100 2026-03-10T11:40:20.921 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/100 2026-03-10T11:40:20.922 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/100 2026-03-10T11:40:20.923 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/100 2026-03-10T11:40:20.927 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/100 2026-03-10T11:40:20.932 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 57/100 2026-03-10T11:40:20.933 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/100 2026-03-10T11:40:20.936 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/100 2026-03-10T11:40:20.940 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 58/100 2026-03-10T11:40:20.941 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/100 2026-03-10T11:40:20.943 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/100 2026-03-10T11:40:20.945 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 59/100 2026-03-10T11:40:20.946 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/100 2026-03-10T11:40:20.947 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 60/100 2026-03-10T11:40:20.949 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/100 2026-03-10T11:40:20.950 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 61/100 2026-03-10T11:40:20.953 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 62/100 2026-03-10T11:40:20.958 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 63/100 2026-03-10T11:40:20.963 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 64/100 2026-03-10T11:40:20.968 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 65/100 2026-03-10T11:40:20.969 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/100 2026-03-10T11:40:20.969 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T11:40:20.969 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T11:40:20.969 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:20.970 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/100 2026-03-10T11:40:20.973 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 66/100 2026-03-10T11:40:20.977 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/100 2026-03-10T11:40:20.978 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/100 2026-03-10T11:40:20.980 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 67/100 2026-03-10T11:40:20.981 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/100 2026-03-10T11:40:20.983 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/100 2026-03-10T11:40:20.983 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 68/100 2026-03-10T11:40:20.986 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/100 2026-03-10T11:40:20.986 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 69/100 2026-03-10T11:40:20.988 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/100 2026-03-10T11:40:20.991 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/100 2026-03-10T11:40:20.992 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 70/100 2026-03-10T11:40:20.993 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 57/100 2026-03-10T11:40:20.996 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 71/100 2026-03-10T11:40:21.000 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 72/100 2026-03-10T11:40:21.001 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 58/100 2026-03-10T11:40:21.005 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 59/100 2026-03-10T11:40:21.007 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 60/100 2026-03-10T11:40:21.008 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 73/100 2026-03-10T11:40:21.009 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 61/100 2026-03-10T11:40:21.012 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 62/100 2026-03-10T11:40:21.014 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 74/100 2026-03-10T11:40:21.017 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 63/100 2026-03-10T11:40:21.017 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 75/100 2026-03-10T11:40:21.021 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 76/100 2026-03-10T11:40:21.021 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 64/100 2026-03-10T11:40:21.023 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 77/100 2026-03-10T11:40:21.026 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 65/100 2026-03-10T11:40:21.029 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 78/100 2026-03-10T11:40:21.030 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 66/100 2026-03-10T11:40:21.033 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 79/100 2026-03-10T11:40:21.036 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 67/100 2026-03-10T11:40:21.039 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 68/100 2026-03-10T11:40:21.041 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 69/100 2026-03-10T11:40:21.046 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 70/100 2026-03-10T11:40:21.050 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 71/100 2026-03-10T11:40:21.053 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 72/100 2026-03-10T11:40:21.054 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 80/100 2026-03-10T11:40:21.054 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-10T11:40:21.054 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:21.061 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 73/100 2026-03-10T11:40:21.061 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 80/100 2026-03-10T11:40:21.066 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 74/100 2026-03-10T11:40:21.069 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 75/100 2026-03-10T11:40:21.072 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 76/100 2026-03-10T11:40:21.073 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 77/100 2026-03-10T11:40:21.078 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 78/100 2026-03-10T11:40:21.082 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 79/100 2026-03-10T11:40:21.088 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 80/100 2026-03-10T11:40:21.088 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 81/100 2026-03-10T11:40:21.100 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 81/100 2026-03-10T11:40:21.100 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 80/100 2026-03-10T11:40:21.101 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-10T11:40:21.101 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:21.105 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 82/100 2026-03-10T11:40:21.107 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 80/100 2026-03-10T11:40:21.108 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 83/100 2026-03-10T11:40:21.110 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 84/100 2026-03-10T11:40:21.111 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 85/100 2026-03-10T11:40:21.134 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 80/100 2026-03-10T11:40:21.134 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 81/100 2026-03-10T11:40:21.146 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 81/100 2026-03-10T11:40:21.150 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 82/100 2026-03-10T11:40:21.153 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 83/100 2026-03-10T11:40:21.155 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 84/100 2026-03-10T11:40:21.155 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 85/100 2026-03-10T11:40:26.346 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 85/100 2026-03-10T11:40:26.346 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /sys 2026-03-10T11:40:26.346 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /proc 2026-03-10T11:40:26.346 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /mnt 2026-03-10T11:40:26.346 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /var/tmp 2026-03-10T11:40:26.346 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /home 2026-03-10T11:40:26.346 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /root 2026-03-10T11:40:26.346 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /tmp 2026-03-10T11:40:26.346 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:26.354 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 86/100 2026-03-10T11:40:26.373 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 87/100 2026-03-10T11:40:26.373 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 87/100 2026-03-10T11:40:26.380 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 87/100 2026-03-10T11:40:26.381 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 85/100 2026-03-10T11:40:26.381 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-03-10T11:40:26.381 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-03-10T11:40:26.381 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-03-10T11:40:26.381 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-03-10T11:40:26.381 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-03-10T11:40:26.381 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-03-10T11:40:26.381 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-03-10T11:40:26.381 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:26.384 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 88/100 2026-03-10T11:40:26.387 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 89/100 2026-03-10T11:40:26.390 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 90/100 2026-03-10T11:40:26.390 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 86/100 2026-03-10T11:40:26.392 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 91/100 2026-03-10T11:40:26.392 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 92/100 2026-03-10T11:40:26.405 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 92/100 2026-03-10T11:40:26.407 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 93/100 2026-03-10T11:40:26.408 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 87/100 2026-03-10T11:40:26.408 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 87/100 2026-03-10T11:40:26.410 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 94/100 2026-03-10T11:40:26.414 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 95/100 2026-03-10T11:40:26.416 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 87/100 2026-03-10T11:40:26.416 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 96/100 2026-03-10T11:40:26.419 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 88/100 2026-03-10T11:40:26.421 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 89/100 2026-03-10T11:40:26.422 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 97/100 2026-03-10T11:40:26.423 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 90/100 2026-03-10T11:40:26.425 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 91/100 2026-03-10T11:40:26.425 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 92/100 2026-03-10T11:40:26.430 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 98/100 2026-03-10T11:40:26.434 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 99/100 2026-03-10T11:40:26.434 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 100/100 2026-03-10T11:40:26.438 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 92/100 2026-03-10T11:40:26.440 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 93/100 2026-03-10T11:40:26.443 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 94/100 2026-03-10T11:40:26.446 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 95/100 2026-03-10T11:40:26.448 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 96/100 2026-03-10T11:40:26.454 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 97/100 2026-03-10T11:40:26.461 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 98/100 2026-03-10T11:40:26.465 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 99/100 2026-03-10T11:40:26.465 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 100/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 100/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/100 2026-03-10T11:40:26.540 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 73/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ply-3.11-14.el9.noarch 74/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 75/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 76/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 78/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 79/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 80/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 81/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 82/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 83/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 84/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 85/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 87/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 88/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 89/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 90/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 91/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 92/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 93/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 94/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 95/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 96/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 97/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 98/100 2026-03-10T11:40:26.541 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 99/100 2026-03-10T11:40:26.560 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 100/100 2026-03-10T11:40:26.560 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/100 2026-03-10T11:40:26.560 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/100 2026-03-10T11:40:26.560 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/100 2026-03-10T11:40:26.560 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/100 2026-03-10T11:40:26.560 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/100 2026-03-10T11:40:26.560 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/100 2026-03-10T11:40:26.560 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/100 2026-03-10T11:40:26.560 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/100 2026-03-10T11:40:26.560 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/100 2026-03-10T11:40:26.562 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/100 2026-03-10T11:40:26.562 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/100 2026-03-10T11:40:26.562 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/100 2026-03-10T11:40:26.562 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/100 2026-03-10T11:40:26.562 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/100 2026-03-10T11:40:26.562 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/100 2026-03-10T11:40:26.562 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/100 2026-03-10T11:40:26.562 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/100 2026-03-10T11:40:26.562 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/100 2026-03-10T11:40:26.562 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/100 2026-03-10T11:40:26.562 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 73/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.noarch 74/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 75/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 76/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 78/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 79/100 2026-03-10T11:40:26.563 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 80/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 81/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 82/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 83/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 84/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 85/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 87/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 88/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 89/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 90/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 91/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 92/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 93/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 94/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 95/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 96/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 97/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 98/100 2026-03-10T11:40:26.564 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 99/100 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 100/100 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T11:40:26.615 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-10T11:40:26.616 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:26.617 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 100/100 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T11:40:26.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:26.642 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:26.806 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout:Remove 1 Package 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 775 k 2026-03-10T11:40:26.807 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T11:40:26.809 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T11:40:26.809 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T11:40:26.810 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T11:40:26.810 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T11:40:26.825 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T11:40:26.825 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T11:40:26.836 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:26.836 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:26.836 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T11:40:26.836 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:26.836 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T11:40:26.836 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-10T11:40:26.836 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:26.836 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T11:40:26.836 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:26.836 INFO:teuthology.orchestra.run.vm06.stdout:Remove 1 Package 2026-03-10T11:40:26.836 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:26.837 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 775 k 2026-03-10T11:40:26.837 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T11:40:26.838 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T11:40:26.838 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T11:40:26.839 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T11:40:26.840 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T11:40:26.855 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T11:40:26.855 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T11:40:26.924 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T11:40:26.955 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T11:40:26.968 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T11:40:26.968 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:26.968 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T11:40:26.968 INFO:teuthology.orchestra.run.vm09.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:26.968 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:26.968 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:27.007 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T11:40:27.007 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:27.007 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T11:40:27.007 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T11:40:27.007 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:27.007 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:27.131 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-immutable-object-cache 2026-03-10T11:40:27.132 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:27.135 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:27.135 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:27.135 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:27.177 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-immutable-object-cache 2026-03-10T11:40:27.177 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:27.180 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:27.181 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:27.181 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:27.291 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-mgr 2026-03-10T11:40:27.292 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:27.295 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:27.295 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:27.295 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:27.337 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr 2026-03-10T11:40:27.337 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:27.340 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:27.341 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:27.341 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:27.451 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-mgr-dashboard 2026-03-10T11:40:27.451 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:27.454 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:27.454 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:27.454 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:27.498 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-dashboard 2026-03-10T11:40:27.498 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:27.501 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:27.502 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:27.502 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:27.612 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-10T11:40:27.612 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:27.615 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:27.615 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:27.616 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:27.658 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-10T11:40:27.658 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:27.661 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:27.662 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:27.662 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:27.771 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-mgr-rook 2026-03-10T11:40:27.771 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:27.774 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:27.775 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:27.775 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:27.816 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-rook 2026-03-10T11:40:27.816 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:27.819 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:27.819 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:27.819 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:27.927 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-mgr-cephadm 2026-03-10T11:40:27.928 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:27.930 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:27.931 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:27.931 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:27.971 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-cephadm 2026-03-10T11:40:27.971 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:27.974 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:27.975 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:27.975 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout:Remove 1 Package 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 3.6 M 2026-03-10T11:40:28.096 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T11:40:28.098 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T11:40:28.098 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T11:40:28.107 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T11:40:28.107 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T11:40:28.131 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout:Remove 1 Package 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 3.6 M 2026-03-10T11:40:28.139 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T11:40:28.141 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T11:40:28.141 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T11:40:28.145 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T11:40:28.150 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T11:40:28.150 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T11:40:28.173 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T11:40:28.188 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T11:40:28.208 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T11:40:28.252 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T11:40:28.264 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T11:40:28.264 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:28.264 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T11:40:28.264 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:28.264 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:28.264 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:28.295 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T11:40:28.295 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:28.295 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T11:40:28.295 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:28.295 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:28.295 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:28.464 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-volume 2026-03-10T11:40:28.464 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:28.467 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:28.468 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:28.468 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:28.491 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-volume 2026-03-10T11:40:28.492 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:28.494 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:28.495 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:28.495 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:28.642 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 610 k 2026-03-10T11:40:28.643 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T11:40:28.645 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T11:40:28.645 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T11:40:28.655 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T11:40:28.656 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repo Size 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout:Removing dependent packages: 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:28.664 INFO:teuthology.orchestra.run.vm09.stdout:Remove 2 Packages 2026-03-10T11:40:28.665 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:28.665 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 610 k 2026-03-10T11:40:28.665 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T11:40:28.666 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T11:40:28.666 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T11:40:28.676 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T11:40:28.676 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T11:40:28.679 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T11:40:28.681 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:28.694 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T11:40:28.700 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T11:40:28.702 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:28.715 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T11:40:28.759 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T11:40:28.759 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:28.800 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T11:40:28.800 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T11:40:28.801 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T11:40:28.801 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:28.801 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T11:40:28.801 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:28.801 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:28.801 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:28.801 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:28.861 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T11:40:28.861 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:28.861 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T11:40:28.861 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:28.861 INFO:teuthology.orchestra.run.vm09.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:28.861 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:28.861 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:28.982 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout:Remove 3 Packages 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 3.7 M 2026-03-10T11:40:28.983 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T11:40:28.985 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T11:40:28.985 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T11:40:29.001 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T11:40:29.002 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T11:40:29.031 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T11:40:29.033 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T11:40:29.035 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T11:40:29.035 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T11:40:29.063 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:29.063 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:29.063 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repo Size 2026-03-10T11:40:29.063 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:29.063 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T11:40:29.063 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-10T11:40:29.063 INFO:teuthology.orchestra.run.vm09.stdout:Removing dependent packages: 2026-03-10T11:40:29.063 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-10T11:40:29.063 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T11:40:29.063 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-10T11:40:29.064 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:29.064 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T11:40:29.064 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:29.064 INFO:teuthology.orchestra.run.vm09.stdout:Remove 3 Packages 2026-03-10T11:40:29.064 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:29.064 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 3.7 M 2026-03-10T11:40:29.064 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T11:40:29.065 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T11:40:29.065 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T11:40:29.081 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T11:40:29.081 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T11:40:29.092 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T11:40:29.092 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T11:40:29.092 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T11:40:29.112 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T11:40:29.114 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T11:40:29.115 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T11:40:29.115 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T11:40:29.130 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T11:40:29.130 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:29.130 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T11:40:29.130 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.130 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.130 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.130 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:29.130 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:29.173 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T11:40:29.173 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T11:40:29.173 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T11:40:29.225 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T11:40:29.225 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:29.225 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T11:40:29.225 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.225 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.225 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.225 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:29.225 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:29.291 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: libcephfs-devel 2026-03-10T11:40:29.291 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:29.294 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:29.295 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:29.295 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:29.384 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: libcephfs-devel 2026-03-10T11:40:29.384 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:29.387 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:29.387 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:29.387 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:29.462 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout:Remove 20 Packages 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:29.464 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 79 M 2026-03-10T11:40:29.465 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T11:40:29.468 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T11:40:29.468 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T11:40:29.489 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T11:40:29.490 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T11:40:29.529 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T11:40:29.531 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-10T11:40:29.533 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-10T11:40:29.536 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-10T11:40:29.536 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T11:40:29.549 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T11:40:29.551 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-10T11:40:29.554 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:29.554 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout:Removing dependent packages: 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-10T11:40:29.555 INFO:teuthology.orchestra.run.vm09.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm09.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm09.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm09.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm09.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm09.stdout:Remove 20 Packages 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 79 M 2026-03-10T11:40:29.556 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T11:40:29.557 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-10T11:40:29.559 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T11:40:29.559 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T11:40:29.560 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-10T11:40:29.560 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T11:40:29.574 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T11:40:29.574 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T11:40:29.574 INFO:teuthology.orchestra.run.vm06.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-10T11:40:29.574 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:29.580 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T11:40:29.580 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T11:40:29.590 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T11:40:29.593 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-10T11:40:29.596 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-10T11:40:29.601 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-10T11:40:29.604 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-10T11:40:29.606 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-10T11:40:29.608 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-10T11:40:29.610 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-10T11:40:29.612 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-10T11:40:29.618 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T11:40:29.620 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-10T11:40:29.622 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-10T11:40:29.625 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-10T11:40:29.625 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T11:40:29.625 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T11:40:29.638 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T11:40:29.640 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-10T11:40:29.642 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-10T11:40:29.643 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T11:40:29.645 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-10T11:40:29.648 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-10T11:40:29.648 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T11:40:29.661 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T11:40:29.661 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T11:40:29.661 INFO:teuthology.orchestra.run.vm09.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-10T11:40:29.661 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:29.675 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T11:40:29.677 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-10T11:40:29.681 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-10T11:40:29.684 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-10T11:40:29.687 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-10T11:40:29.689 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-10T11:40:29.690 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-10T11:40:29.692 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-10T11:40:29.694 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-10T11:40:29.696 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-10T11:40:29.709 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.734 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T11:40:29.735 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T11:40:29.735 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.735 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.735 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.735 INFO:teuthology.orchestra.run.vm06.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-10T11:40:29.735 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.735 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.735 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T11:40:29.735 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T11:40:29.735 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T11:40:29.735 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:29.772 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T11:40:29.772 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-10T11:40:29.772 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-10T11:40:29.772 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-10T11:40:29.772 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-10T11:40:29.772 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-10T11:40:29.772 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-10T11:40:29.772 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T11:40:29.772 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-10T11:40:29.772 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-10T11:40:29.773 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T11:40:29.773 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-10T11:40:29.773 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-10T11:40:29.773 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-10T11:40:29.773 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-10T11:40:29.773 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-10T11:40:29.773 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-10T11:40:29.773 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-10T11:40:29.773 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-10T11:40:29.773 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-10T11:40:29.814 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-10T11:40:29.814 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:29.814 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T11:40:29.814 INFO:teuthology.orchestra.run.vm09.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T11:40:29.814 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T11:40:29.814 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T11:40:29.815 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:29.933 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: librbd1 2026-03-10T11:40:29.933 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:29.936 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:29.936 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:29.936 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:30.008 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: librbd1 2026-03-10T11:40:30.008 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:30.010 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:30.010 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:30.010 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:30.144 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rados 2026-03-10T11:40:30.144 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:30.146 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:30.146 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:30.146 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:30.178 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: python3-rados 2026-03-10T11:40:30.178 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:30.180 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:30.180 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:30.180 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:30.304 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rgw 2026-03-10T11:40:30.304 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:30.306 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:30.307 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:30.307 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:30.335 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: python3-rgw 2026-03-10T11:40:30.335 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:30.337 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:30.338 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:30.338 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:30.464 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-cephfs 2026-03-10T11:40:30.464 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:30.466 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:30.466 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:30.466 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:30.491 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: python3-cephfs 2026-03-10T11:40:30.491 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:30.493 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:30.493 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:30.494 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:30.625 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rbd 2026-03-10T11:40:30.625 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:30.627 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:30.628 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:30.628 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:30.646 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: python3-rbd 2026-03-10T11:40:30.646 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:30.648 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:30.649 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:30.649 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:30.785 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-fuse 2026-03-10T11:40:30.785 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:30.787 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:30.788 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:30.788 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:30.801 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: rbd-fuse 2026-03-10T11:40:30.801 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:30.803 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:30.803 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:30.803 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:30.940 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-mirror 2026-03-10T11:40:30.940 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:30.942 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:30.942 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:30.942 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:30.955 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: rbd-mirror 2026-03-10T11:40:30.955 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:30.957 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:30.957 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:30.957 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:31.099 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-nbd 2026-03-10T11:40:31.099 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T11:40:31.101 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T11:40:31.102 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T11:40:31.102 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T11:40:31.111 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: rbd-nbd 2026-03-10T11:40:31.111 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T11:40:31.113 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T11:40:31.113 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T11:40:31.113 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T11:40:31.122 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-03-10T11:40:31.133 DEBUG:teuthology.orchestra.run.vm09:> sudo yum clean all 2026-03-10T11:40:31.245 INFO:teuthology.orchestra.run.vm06.stdout:56 files removed 2026-03-10T11:40:31.257 INFO:teuthology.orchestra.run.vm09.stdout:56 files removed 2026-03-10T11:40:31.265 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T11:40:31.277 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T11:40:31.287 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean expire-cache 2026-03-10T11:40:31.300 DEBUG:teuthology.orchestra.run.vm09:> sudo yum clean expire-cache 2026-03-10T11:40:31.432 INFO:teuthology.orchestra.run.vm06.stdout:Cache was expired 2026-03-10T11:40:31.432 INFO:teuthology.orchestra.run.vm06.stdout:0 files removed 2026-03-10T11:40:31.448 INFO:teuthology.orchestra.run.vm09.stdout:Cache was expired 2026-03-10T11:40:31.448 INFO:teuthology.orchestra.run.vm09.stdout:0 files removed 2026-03-10T11:40:31.450 DEBUG:teuthology.parallel:result is None 2026-03-10T11:40:31.466 DEBUG:teuthology.parallel:result is None 2026-03-10T11:40:31.466 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm06.local 2026-03-10T11:40:31.466 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm09.local 2026-03-10T11:40:31.466 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T11:40:31.467 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T11:40:31.490 DEBUG:teuthology.orchestra.run.vm09:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-10T11:40:31.490 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-10T11:40:31.553 DEBUG:teuthology.parallel:result is None 2026-03-10T11:40:31.553 DEBUG:teuthology.parallel:result is None 2026-03-10T11:40:31.553 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-10T11:40:31.555 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-10T11:40:31.555 DEBUG:teuthology.orchestra.run.vm06:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T11:40:31.595 DEBUG:teuthology.orchestra.run.vm09:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T11:40:31.607 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-10T11:40:31.608 INFO:teuthology.orchestra.run.vm09.stderr:bash: line 1: ntpq: command not found 2026-03-10T11:40:31.609 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T11:40:31.610 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-10T11:40:31.610 INFO:teuthology.orchestra.run.vm06.stdout:^- fa.gnudb.org 2 6 377 48 +565us[ +549us] +/- 47ms 2026-03-10T11:40:31.610 INFO:teuthology.orchestra.run.vm06.stdout:^+ 47.ip-51-75-67.eu 4 6 377 46 -1025us[-1025us] +/- 16ms 2026-03-10T11:40:31.610 INFO:teuthology.orchestra.run.vm06.stdout:^* node-1.infogral.is 2 6 377 46 +22us[+5602ns] +/- 14ms 2026-03-10T11:40:31.610 INFO:teuthology.orchestra.run.vm06.stdout:^+ ntp2.uni-ulm.de 2 6 377 47 +1027us[+1011us] +/- 15ms 2026-03-10T11:40:31.611 INFO:teuthology.orchestra.run.vm09.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T11:40:31.611 INFO:teuthology.orchestra.run.vm09.stdout:=============================================================================== 2026-03-10T11:40:31.611 INFO:teuthology.orchestra.run.vm09.stdout:^* node-1.infogral.is 2 6 377 45 -25us[ -44us] +/- 14ms 2026-03-10T11:40:31.611 INFO:teuthology.orchestra.run.vm09.stdout:^+ ntp2.uni-ulm.de 2 6 377 48 +898us[ +878us] +/- 15ms 2026-03-10T11:40:31.611 INFO:teuthology.orchestra.run.vm09.stdout:^- fa.gnudb.org 2 6 377 49 +360us[ +341us] +/- 47ms 2026-03-10T11:40:31.611 INFO:teuthology.orchestra.run.vm09.stdout:^+ 47.ip-51-75-67.eu 4 6 377 48 -1095us[-1114us] +/- 16ms 2026-03-10T11:40:31.611 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-10T11:40:31.614 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-10T11:40:31.614 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-10T11:40:31.616 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-10T11:40:31.617 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-10T11:40:31.619 INFO:teuthology.task.internal:Duration was 889.814171 seconds 2026-03-10T11:40:31.619 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-10T11:40:31.621 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-10T11:40:31.621 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T11:40:31.652 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T11:40:31.684 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T11:40:31.686 INFO:teuthology.orchestra.run.vm09.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T11:40:32.058 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-10T11:40:32.058 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm06.local 2026-03-10T11:40:32.058 DEBUG:teuthology.orchestra.run.vm06:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T11:40:32.120 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm09.local 2026-03-10T11:40:32.120 DEBUG:teuthology.orchestra.run.vm09:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T11:40:32.143 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-10T11:40:32.143 DEBUG:teuthology.orchestra.run.vm06:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T11:40:32.161 DEBUG:teuthology.orchestra.run.vm09:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T11:40:32.607 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-10T11:40:32.607 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T11:40:32.609 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T11:40:32.629 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T11:40:32.629 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T11:40:32.629 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T11:40:32.630 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T11:40:32.630 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T11:40:32.631 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T11:40:32.631 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T11:40:32.631 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T11:40:32.632 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T11:40:32.632 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz/home/ubuntu/cephtest/archive/syslog/journalctl.log: 2026-03-10T11:40:32.748 INFO:teuthology.orchestra.run.vm06.stderr: 97.7% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T11:40:32.761 INFO:teuthology.orchestra.run.vm09.stderr: 98.1% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T11:40:32.763 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-10T11:40:32.765 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-10T11:40:32.765 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T11:40:32.817 DEBUG:teuthology.orchestra.run.vm09:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T11:40:32.839 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-10T11:40:32.842 DEBUG:teuthology.orchestra.run.vm06:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T11:40:32.859 DEBUG:teuthology.orchestra.run.vm09:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T11:40:32.882 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = core 2026-03-10T11:40:32.905 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = core 2026-03-10T11:40:32.916 DEBUG:teuthology.orchestra.run.vm06:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T11:40:32.948 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T11:40:32.949 DEBUG:teuthology.orchestra.run.vm09:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T11:40:32.971 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T11:40:32.971 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-10T11:40:32.974 INFO:teuthology.task.internal:Transferring archived files... 2026-03-10T11:40:32.974 DEBUG:teuthology.misc:Transferring archived files from vm06:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1016/remote/vm06 2026-03-10T11:40:32.974 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T11:40:33.017 DEBUG:teuthology.misc:Transferring archived files from vm09:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1016/remote/vm09 2026-03-10T11:40:33.017 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T11:40:33.044 INFO:teuthology.task.internal:Removing archive directory... 2026-03-10T11:40:33.044 DEBUG:teuthology.orchestra.run.vm06:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T11:40:33.058 DEBUG:teuthology.orchestra.run.vm09:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T11:40:33.099 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-10T11:40:33.102 INFO:teuthology.task.internal:Not uploading archives. 2026-03-10T11:40:33.102 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-10T11:40:33.104 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-10T11:40:33.104 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T11:40:33.113 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T11:40:33.128 INFO:teuthology.orchestra.run.vm06.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 11:40 /home/ubuntu/cephtest 2026-03-10T11:40:33.155 INFO:teuthology.orchestra.run.vm09.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 11:40 /home/ubuntu/cephtest 2026-03-10T11:40:33.156 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-10T11:40:33.163 INFO:teuthology.run:Summary data: description: orch/cephadm/with-work/{0-distro/centos_9.stream fixed-2 mode/packaged mon_election/classic msgr/async start tasks/rados_python} duration: 889.8141708374023 flavor: default owner: kyr success: true 2026-03-10T11:40:33.163 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T11:40:33.180 INFO:teuthology.run:pass